var/home/core/zuul-output/0000755000175000017500000000000015070411732014525 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070444603015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006751001015070444571017704 0ustar rootrootOct 05 06:52:44 crc systemd[1]: Starting Kubernetes Kubelet... Oct 05 06:52:45 crc restorecon[4795]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:52:45 crc restorecon[4795]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 05 06:52:46 crc kubenswrapper[4935]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 06:52:46 crc kubenswrapper[4935]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 05 06:52:46 crc kubenswrapper[4935]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 06:52:46 crc kubenswrapper[4935]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 06:52:46 crc kubenswrapper[4935]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 05 06:52:46 crc kubenswrapper[4935]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.518174 4935 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522201 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522247 4935 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522254 4935 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522260 4935 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522266 4935 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522272 4935 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522279 4935 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522286 4935 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522292 4935 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522297 4935 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522302 4935 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522315 4935 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522321 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522327 4935 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522334 4935 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522340 4935 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522345 4935 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522350 4935 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522357 4935 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522364 4935 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522369 4935 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522421 4935 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522429 4935 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522436 4935 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522441 4935 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522446 4935 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522451 4935 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522456 4935 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522461 4935 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522467 4935 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522471 4935 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522476 4935 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522481 4935 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522486 4935 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522491 4935 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522496 4935 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522500 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522505 4935 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522510 4935 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522515 4935 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522521 4935 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522526 4935 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522530 4935 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522536 4935 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522543 4935 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522549 4935 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522555 4935 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522559 4935 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522564 4935 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522569 4935 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522574 4935 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522579 4935 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522584 4935 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522590 4935 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522596 4935 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522603 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522609 4935 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522618 4935 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522625 4935 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522630 4935 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522636 4935 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522641 4935 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522647 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522652 4935 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522657 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522664 4935 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522670 4935 feature_gate.go:330] unrecognized feature gate: Example Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522678 4935 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522686 4935 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522693 4935 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.522700 4935 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523769 4935 flags.go:64] FLAG: --address="0.0.0.0" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523792 4935 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523807 4935 flags.go:64] FLAG: --anonymous-auth="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523817 4935 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523825 4935 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523833 4935 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523843 4935 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523853 4935 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523862 4935 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523869 4935 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523878 4935 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523885 4935 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523921 4935 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523929 4935 flags.go:64] FLAG: --cgroup-root="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523936 4935 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523943 4935 flags.go:64] FLAG: --client-ca-file="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523950 4935 flags.go:64] FLAG: --cloud-config="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523957 4935 flags.go:64] FLAG: --cloud-provider="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523964 4935 flags.go:64] FLAG: --cluster-dns="[]" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523972 4935 flags.go:64] FLAG: --cluster-domain="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523978 4935 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523986 4935 flags.go:64] FLAG: --config-dir="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.523992 4935 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524000 4935 flags.go:64] FLAG: --container-log-max-files="5" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524016 4935 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524023 4935 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524031 4935 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524038 4935 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524046 4935 flags.go:64] FLAG: --contention-profiling="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524053 4935 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524060 4935 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524069 4935 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524076 4935 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524085 4935 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524092 4935 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524100 4935 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524108 4935 flags.go:64] FLAG: --enable-load-reader="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524115 4935 flags.go:64] FLAG: --enable-server="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524122 4935 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524135 4935 flags.go:64] FLAG: --event-burst="100" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524144 4935 flags.go:64] FLAG: --event-qps="50" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524152 4935 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524159 4935 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524166 4935 flags.go:64] FLAG: --eviction-hard="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524176 4935 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524183 4935 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524190 4935 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524197 4935 flags.go:64] FLAG: --eviction-soft="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524205 4935 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524211 4935 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524218 4935 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524225 4935 flags.go:64] FLAG: --experimental-mounter-path="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524232 4935 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524239 4935 flags.go:64] FLAG: --fail-swap-on="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524246 4935 flags.go:64] FLAG: --feature-gates="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524254 4935 flags.go:64] FLAG: --file-check-frequency="20s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524262 4935 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524270 4935 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524277 4935 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524284 4935 flags.go:64] FLAG: --healthz-port="10248" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524292 4935 flags.go:64] FLAG: --help="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524298 4935 flags.go:64] FLAG: --hostname-override="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524305 4935 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524314 4935 flags.go:64] FLAG: --http-check-frequency="20s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524321 4935 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524328 4935 flags.go:64] FLAG: --image-credential-provider-config="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524335 4935 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524342 4935 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524349 4935 flags.go:64] FLAG: --image-service-endpoint="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524355 4935 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524362 4935 flags.go:64] FLAG: --kube-api-burst="100" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524370 4935 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524378 4935 flags.go:64] FLAG: --kube-api-qps="50" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524386 4935 flags.go:64] FLAG: --kube-reserved="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524393 4935 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524399 4935 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524407 4935 flags.go:64] FLAG: --kubelet-cgroups="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524414 4935 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524421 4935 flags.go:64] FLAG: --lock-file="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524427 4935 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524436 4935 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524443 4935 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524454 4935 flags.go:64] FLAG: --log-json-split-stream="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524463 4935 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524470 4935 flags.go:64] FLAG: --log-text-split-stream="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524477 4935 flags.go:64] FLAG: --logging-format="text" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524484 4935 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524491 4935 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524498 4935 flags.go:64] FLAG: --manifest-url="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524505 4935 flags.go:64] FLAG: --manifest-url-header="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524514 4935 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524521 4935 flags.go:64] FLAG: --max-open-files="1000000" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524530 4935 flags.go:64] FLAG: --max-pods="110" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524537 4935 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524544 4935 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524553 4935 flags.go:64] FLAG: --memory-manager-policy="None" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524560 4935 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524568 4935 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524575 4935 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524582 4935 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524598 4935 flags.go:64] FLAG: --node-status-max-images="50" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524605 4935 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524612 4935 flags.go:64] FLAG: --oom-score-adj="-999" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524620 4935 flags.go:64] FLAG: --pod-cidr="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524627 4935 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524637 4935 flags.go:64] FLAG: --pod-manifest-path="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524645 4935 flags.go:64] FLAG: --pod-max-pids="-1" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524652 4935 flags.go:64] FLAG: --pods-per-core="0" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524659 4935 flags.go:64] FLAG: --port="10250" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524667 4935 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524673 4935 flags.go:64] FLAG: --provider-id="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524680 4935 flags.go:64] FLAG: --qos-reserved="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524687 4935 flags.go:64] FLAG: --read-only-port="10255" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524694 4935 flags.go:64] FLAG: --register-node="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524702 4935 flags.go:64] FLAG: --register-schedulable="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524709 4935 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524723 4935 flags.go:64] FLAG: --registry-burst="10" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524730 4935 flags.go:64] FLAG: --registry-qps="5" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524738 4935 flags.go:64] FLAG: --reserved-cpus="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524745 4935 flags.go:64] FLAG: --reserved-memory="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524753 4935 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524760 4935 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524768 4935 flags.go:64] FLAG: --rotate-certificates="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524775 4935 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524781 4935 flags.go:64] FLAG: --runonce="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524788 4935 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524796 4935 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524804 4935 flags.go:64] FLAG: --seccomp-default="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524810 4935 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524817 4935 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524824 4935 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524831 4935 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524838 4935 flags.go:64] FLAG: --storage-driver-password="root" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524845 4935 flags.go:64] FLAG: --storage-driver-secure="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524852 4935 flags.go:64] FLAG: --storage-driver-table="stats" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524859 4935 flags.go:64] FLAG: --storage-driver-user="root" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524866 4935 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524874 4935 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524881 4935 flags.go:64] FLAG: --system-cgroups="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524888 4935 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524921 4935 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524928 4935 flags.go:64] FLAG: --tls-cert-file="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524934 4935 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524943 4935 flags.go:64] FLAG: --tls-min-version="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524950 4935 flags.go:64] FLAG: --tls-private-key-file="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524956 4935 flags.go:64] FLAG: --topology-manager-policy="none" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524964 4935 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524971 4935 flags.go:64] FLAG: --topology-manager-scope="container" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524978 4935 flags.go:64] FLAG: --v="2" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524987 4935 flags.go:64] FLAG: --version="false" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.524996 4935 flags.go:64] FLAG: --vmodule="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.525004 4935 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.525013 4935 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525185 4935 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525195 4935 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525203 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525211 4935 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525218 4935 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525225 4935 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525234 4935 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525242 4935 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525249 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525256 4935 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525263 4935 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525269 4935 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525276 4935 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525284 4935 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525291 4935 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525302 4935 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525308 4935 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525314 4935 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525321 4935 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525327 4935 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525333 4935 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525341 4935 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525349 4935 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525355 4935 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525362 4935 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525369 4935 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525376 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525382 4935 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525389 4935 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525395 4935 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525401 4935 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525407 4935 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525413 4935 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525419 4935 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525425 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525434 4935 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525440 4935 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525447 4935 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525453 4935 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525459 4935 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525467 4935 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525475 4935 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525482 4935 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525489 4935 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525497 4935 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525504 4935 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525511 4935 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525520 4935 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525526 4935 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525532 4935 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525542 4935 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525548 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525554 4935 feature_gate.go:330] unrecognized feature gate: Example Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525605 4935 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525613 4935 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525619 4935 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525625 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525631 4935 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525637 4935 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525643 4935 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525650 4935 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525655 4935 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525661 4935 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525667 4935 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525673 4935 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525679 4935 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525685 4935 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525691 4935 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525697 4935 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525704 4935 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.525711 4935 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.526775 4935 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.540039 4935 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.540125 4935 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540274 4935 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540292 4935 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540305 4935 feature_gate.go:330] unrecognized feature gate: Example Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540315 4935 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540325 4935 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540334 4935 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540343 4935 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540352 4935 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540360 4935 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.540475 4935 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541320 4935 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541355 4935 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541369 4935 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541381 4935 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541394 4935 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541408 4935 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541423 4935 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541437 4935 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541450 4935 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541461 4935 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541474 4935 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541485 4935 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541498 4935 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541520 4935 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541529 4935 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541545 4935 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541562 4935 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541574 4935 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541583 4935 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541593 4935 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541606 4935 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541617 4935 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541627 4935 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541637 4935 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541655 4935 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541666 4935 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541676 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541687 4935 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541698 4935 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541707 4935 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541717 4935 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541727 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541739 4935 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541751 4935 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541762 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541865 4935 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541878 4935 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541927 4935 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541939 4935 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541951 4935 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541962 4935 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541973 4935 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541984 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.541999 4935 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542010 4935 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542019 4935 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542029 4935 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542038 4935 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542048 4935 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542057 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542065 4935 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542074 4935 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542086 4935 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542097 4935 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542109 4935 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542125 4935 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542135 4935 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542144 4935 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542153 4935 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542162 4935 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.542172 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.542189 4935 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543482 4935 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543522 4935 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543535 4935 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543544 4935 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543555 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543564 4935 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543572 4935 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543581 4935 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543589 4935 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543598 4935 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543608 4935 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543621 4935 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543635 4935 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543645 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543656 4935 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543667 4935 feature_gate.go:330] unrecognized feature gate: Example Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543676 4935 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543685 4935 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543694 4935 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543705 4935 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543714 4935 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543723 4935 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543731 4935 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543743 4935 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543755 4935 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543766 4935 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543776 4935 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543785 4935 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543794 4935 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543802 4935 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543811 4935 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543821 4935 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543831 4935 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543841 4935 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543852 4935 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543862 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543871 4935 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543881 4935 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543927 4935 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543938 4935 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543949 4935 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543959 4935 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543969 4935 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543978 4935 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543987 4935 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.543995 4935 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544004 4935 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544013 4935 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544026 4935 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544035 4935 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544044 4935 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544053 4935 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544063 4935 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544071 4935 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544080 4935 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544088 4935 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544097 4935 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544106 4935 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544115 4935 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544123 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544132 4935 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544140 4935 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544149 4935 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544157 4935 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544165 4935 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544177 4935 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544186 4935 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544195 4935 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544203 4935 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544212 4935 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.544220 4935 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.544237 4935 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.546156 4935 server.go:940] "Client rotation is on, will bootstrap in background" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.552923 4935 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.553079 4935 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.554955 4935 server.go:997] "Starting client certificate rotation" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.555007 4935 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.556155 4935 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-28 19:50:23.401647294 +0000 UTC Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.556224 4935 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1308h57m36.845425972s for next certificate rotation Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.586523 4935 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.589539 4935 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.611856 4935 log.go:25] "Validated CRI v1 runtime API" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.653328 4935 log.go:25] "Validated CRI v1 image API" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.655504 4935 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.663454 4935 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-05-06-43-26-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.663491 4935 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.681329 4935 manager.go:217] Machine: {Timestamp:2025-10-05 06:52:46.676565042 +0000 UTC m=+0.559191512 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:dddde716-0390-449f-afc3-a07540401a49 BootID:b377934a-1ade-44d6-a53c-ef84970bcbd9 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:66:2e:e8 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:66:2e:e8 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ea:6c:eb Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:42:39:fc Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:61:a8:e9 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2e:17:bb Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:11:40:66 Speed:-1 Mtu:1496} {Name:ens7.44 MacAddress:52:54:00:1a:b9:54 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ea:51:84:c0:1c:f3 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5e:ad:63:b8:9b:9b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.681799 4935 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.682184 4935 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.682825 4935 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.683166 4935 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.683242 4935 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.683625 4935 topology_manager.go:138] "Creating topology manager with none policy" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.683646 4935 container_manager_linux.go:303] "Creating device plugin manager" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.684103 4935 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.684165 4935 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.685605 4935 state_mem.go:36] "Initialized new in-memory state store" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.685807 4935 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.691943 4935 kubelet.go:418] "Attempting to sync node with API server" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.691978 4935 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.692098 4935 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.692120 4935 kubelet.go:324] "Adding apiserver pod source" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.692151 4935 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.698329 4935 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.699630 4935 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.701371 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.701431 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.701532 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.701692 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.702922 4935 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705335 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705385 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705406 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705420 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705450 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705462 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705473 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705489 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705500 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705511 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705525 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.705535 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.706425 4935 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.707158 4935 server.go:1280] "Started kubelet" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.707430 4935 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.707754 4935 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.708024 4935 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:46 crc systemd[1]: Started Kubernetes Kubelet. Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.710106 4935 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.711780 4935 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.711827 4935 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.712284 4935 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.712383 4935 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.712271 4935 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.712471 4935 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.712377 4935 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 20:44:02.522349093 +0000 UTC Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.712529 4935 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2029h51m15.809829934s for next certificate rotation Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.713348 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.713436 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.713644 4935 server.go:460] "Adding debug handlers to kubelet server" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.717110 4935 factory.go:153] Registering CRI-O factory Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.717213 4935 factory.go:221] Registration of the crio container factory successfully Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.717356 4935 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.717389 4935 factory.go:55] Registering systemd factory Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.717456 4935 factory.go:221] Registration of the systemd container factory successfully Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.717504 4935 factory.go:103] Registering Raw factory Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.717505 4935 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.59:6443: connect: connection refused" interval="200ms" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.717542 4935 manager.go:1196] Started watching for new ooms in manager Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.719358 4935 manager.go:319] Starting recovery of all containers Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.720240 4935 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.59:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b85c66d9997e0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-05 06:52:46.707111904 +0000 UTC m=+0.589738384,LastTimestamp:2025-10-05 06:52:46.707111904 +0000 UTC m=+0.589738384,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732303 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732355 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732373 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732391 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732408 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732423 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732436 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732450 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732468 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732483 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732497 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732510 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732527 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732543 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732557 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732574 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732589 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732605 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732619 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732635 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732650 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732693 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732709 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732725 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732740 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732757 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732777 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732792 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732806 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732820 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732838 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732854 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732922 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732943 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732958 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732971 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.732986 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733000 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733015 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733028 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733039 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733051 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733062 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733073 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733084 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733095 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733106 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733119 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733131 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733142 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733155 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733167 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733184 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733198 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733210 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733220 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733233 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733244 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733254 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733265 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733276 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733286 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733298 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733311 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733322 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733332 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733343 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733353 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733365 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733375 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733386 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733397 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733409 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733420 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733432 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733449 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733464 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733479 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733493 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733509 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733524 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733539 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733554 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733567 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733581 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733594 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733608 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733623 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733636 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733649 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733662 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733678 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733693 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733749 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733766 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.733783 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.736773 4935 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.736837 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.736866 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.736963 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737001 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737036 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737067 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737099 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737129 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737177 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737215 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737252 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737286 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737320 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737353 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737392 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737423 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737457 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737488 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737518 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737546 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737578 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737610 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737640 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737668 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737696 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737724 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737800 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737833 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737868 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.737964 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738001 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738035 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738088 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738117 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738145 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738174 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738203 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738231 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738259 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738285 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738314 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738342 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738375 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738406 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738434 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738461 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738491 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738520 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738546 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738573 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738602 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738629 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738656 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738685 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738710 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738741 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738765 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738795 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738822 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738849 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738882 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738947 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.738978 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739005 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739033 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739061 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739091 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739117 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739144 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739174 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739201 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739231 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739258 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739285 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739315 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739341 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739366 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739393 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739424 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739450 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739477 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739505 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739530 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739555 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739581 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739606 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739631 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739657 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739685 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739712 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739738 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739764 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739790 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739818 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739847 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739876 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739934 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739961 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.739991 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740021 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740052 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740079 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740113 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740142 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740170 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740198 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740226 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740253 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740277 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740306 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740339 4935 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740366 4935 reconstruct.go:97] "Volume reconstruction finished" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.740383 4935 reconciler.go:26] "Reconciler: start to sync state" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.747127 4935 manager.go:324] Recovery completed Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.763359 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.765088 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.765123 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.765134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.765800 4935 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.765823 4935 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.765844 4935 state_mem.go:36] "Initialized new in-memory state store" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.772526 4935 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.775804 4935 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.775860 4935 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.775923 4935 kubelet.go:2335] "Starting kubelet main sync loop" Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.776000 4935 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 05 06:52:46 crc kubenswrapper[4935]: W1005 06:52:46.776817 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.776928 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.784529 4935 policy_none.go:49] "None policy: Start" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.786057 4935 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.786105 4935 state_mem.go:35] "Initializing new in-memory state store" Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.813426 4935 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.852800 4935 manager.go:334] "Starting Device Plugin manager" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.852874 4935 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.852911 4935 server.go:79] "Starting device plugin registration server" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.853462 4935 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.853488 4935 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.853657 4935 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.853840 4935 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.853857 4935 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.863555 4935 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.876276 4935 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.876395 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.877934 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.878012 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.878027 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.878233 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.878508 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.878563 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.879707 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.879739 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.879752 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.879881 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.879965 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.880165 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.880424 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.880595 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.880645 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.881991 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.882031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.882042 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.882317 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.882409 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.882436 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.882796 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.883002 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.883083 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.884664 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.884697 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.884711 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.884762 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.884729 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.884841 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.885005 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.885162 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.885207 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.886166 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.886188 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.886198 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.886370 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.886398 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.886589 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.886635 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.886651 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.887195 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.887230 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.887243 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.918451 4935 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.59:6443: connect: connection refused" interval="400ms" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.942907 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.942985 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943075 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943107 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943163 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943182 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943243 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943259 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943307 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943334 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943396 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943415 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943712 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943767 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.943797 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.953819 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.955415 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.955466 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.955484 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:46 crc kubenswrapper[4935]: I1005 06:52:46.955521 4935 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:52:46 crc kubenswrapper[4935]: E1005 06:52:46.956150 4935 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.59:6443: connect: connection refused" node="crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.044732 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.044812 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.044850 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.044881 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.044937 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.044969 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045021 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045026 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045102 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045135 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045178 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045258 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045274 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045203 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045247 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045380 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045384 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045467 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045506 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045492 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045530 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045570 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045773 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045797 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045878 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045826 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.046121 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045961 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.045992 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.046279 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.156714 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.158042 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.158101 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.158155 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.158193 4935 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:52:47 crc kubenswrapper[4935]: E1005 06:52:47.158768 4935 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.59:6443: connect: connection refused" node="crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.213498 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.222450 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.244100 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.254335 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.258760 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:47 crc kubenswrapper[4935]: W1005 06:52:47.261274 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-de7722304541d50c99cec96d2cf2e70541add9ec7913d49bc87e0abc097ab746 WatchSource:0}: Error finding container de7722304541d50c99cec96d2cf2e70541add9ec7913d49bc87e0abc097ab746: Status 404 returned error can't find the container with id de7722304541d50c99cec96d2cf2e70541add9ec7913d49bc87e0abc097ab746 Oct 05 06:52:47 crc kubenswrapper[4935]: W1005 06:52:47.265929 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-6ce49933605c7f4861d70fdf3d887d7b4da0515c17fb69e2f6c51f46cbb2d10c WatchSource:0}: Error finding container 6ce49933605c7f4861d70fdf3d887d7b4da0515c17fb69e2f6c51f46cbb2d10c: Status 404 returned error can't find the container with id 6ce49933605c7f4861d70fdf3d887d7b4da0515c17fb69e2f6c51f46cbb2d10c Oct 05 06:52:47 crc kubenswrapper[4935]: W1005 06:52:47.269664 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7f09be1eb8e66aeff3f0ccbb3d4a7fadb53f24117443370e7a9727f3b41ffc54 WatchSource:0}: Error finding container 7f09be1eb8e66aeff3f0ccbb3d4a7fadb53f24117443370e7a9727f3b41ffc54: Status 404 returned error can't find the container with id 7f09be1eb8e66aeff3f0ccbb3d4a7fadb53f24117443370e7a9727f3b41ffc54 Oct 05 06:52:47 crc kubenswrapper[4935]: W1005 06:52:47.275119 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-eac290d69d9e7adfb6d6a4b5d9d99d05ea99b697c4e1b9086d1df61dbe31b233 WatchSource:0}: Error finding container eac290d69d9e7adfb6d6a4b5d9d99d05ea99b697c4e1b9086d1df61dbe31b233: Status 404 returned error can't find the container with id eac290d69d9e7adfb6d6a4b5d9d99d05ea99b697c4e1b9086d1df61dbe31b233 Oct 05 06:52:47 crc kubenswrapper[4935]: E1005 06:52:47.319552 4935 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.59:6443: connect: connection refused" interval="800ms" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.560370 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.562220 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.562278 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.562293 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.562362 4935 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:52:47 crc kubenswrapper[4935]: E1005 06:52:47.562935 4935 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.59:6443: connect: connection refused" node="crc" Oct 05 06:52:47 crc kubenswrapper[4935]: W1005 06:52:47.683024 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:47 crc kubenswrapper[4935]: E1005 06:52:47.683123 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.710242 4935 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.780288 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"eac290d69d9e7adfb6d6a4b5d9d99d05ea99b697c4e1b9086d1df61dbe31b233"} Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.781646 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7f09be1eb8e66aeff3f0ccbb3d4a7fadb53f24117443370e7a9727f3b41ffc54"} Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.782804 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6ce49933605c7f4861d70fdf3d887d7b4da0515c17fb69e2f6c51f46cbb2d10c"} Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.783990 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"de7722304541d50c99cec96d2cf2e70541add9ec7913d49bc87e0abc097ab746"} Oct 05 06:52:47 crc kubenswrapper[4935]: I1005 06:52:47.785108 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6a6ff173550408bccd9ca4418b7e8d51d41087c996b0723988cb253f3864e30a"} Oct 05 06:52:47 crc kubenswrapper[4935]: W1005 06:52:47.800303 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:47 crc kubenswrapper[4935]: E1005 06:52:47.800383 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:48 crc kubenswrapper[4935]: W1005 06:52:48.087004 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:48 crc kubenswrapper[4935]: E1005 06:52:48.087104 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:48 crc kubenswrapper[4935]: E1005 06:52:48.121216 4935 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.59:6443: connect: connection refused" interval="1.6s" Oct 05 06:52:48 crc kubenswrapper[4935]: W1005 06:52:48.183291 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:48 crc kubenswrapper[4935]: E1005 06:52:48.183403 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.363377 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.365349 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.365429 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.365450 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.365527 4935 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:52:48 crc kubenswrapper[4935]: E1005 06:52:48.366167 4935 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.59:6443: connect: connection refused" node="crc" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.710669 4935 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.792886 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753"} Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.793001 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc"} Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.793022 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21"} Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.793041 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7"} Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.793202 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.794711 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.794750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.794759 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.796198 4935 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6" exitCode=0 Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.796339 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6"} Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.796419 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.798542 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.798599 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.798619 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.799222 4935 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b" exitCode=0 Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.799342 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.799344 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b"} Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.800352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.800379 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.800388 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.801810 4935 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f9374586a384472fd723385827cdc9268e30d02def3740c44e7a4e56abbdd592" exitCode=0 Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.801867 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f9374586a384472fd723385827cdc9268e30d02def3740c44e7a4e56abbdd592"} Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.802040 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.802577 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.803494 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.803576 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.803608 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.803727 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.803746 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.803757 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.804560 4935 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6c71942182bd653c831680f6c9d9ef0b2e58a5747904364e2ce3a6a5d9786e81" exitCode=0 Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.804601 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6c71942182bd653c831680f6c9d9ef0b2e58a5747904364e2ce3a6a5d9786e81"} Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.804723 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.806186 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.806218 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:48 crc kubenswrapper[4935]: I1005 06:52:48.806229 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:49 crc kubenswrapper[4935]: E1005 06:52:49.144005 4935 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.59:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b85c66d9997e0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-05 06:52:46.707111904 +0000 UTC m=+0.589738384,LastTimestamp:2025-10-05 06:52:46.707111904 +0000 UTC m=+0.589738384,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.710694 4935 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:49 crc kubenswrapper[4935]: E1005 06:52:49.722783 4935 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.59:6443: connect: connection refused" interval="3.2s" Oct 05 06:52:49 crc kubenswrapper[4935]: W1005 06:52:49.760584 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:49 crc kubenswrapper[4935]: E1005 06:52:49.760693 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.818157 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.818220 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.818230 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.818241 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.820201 4935 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3ae9cc68210f7d97c95a79f044f979b65a11be9828c9ce22456d497e5cd25975" exitCode=0 Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.820285 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3ae9cc68210f7d97c95a79f044f979b65a11be9828c9ce22456d497e5cd25975"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.820446 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.822226 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.822265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.822286 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.824422 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.824405 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"af3c608e5fb2b5e5244d6cf09e93f739f8c8a6171452e5a621e860de7ad186a2"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.826067 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.826112 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.826125 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.827852 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.827929 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.827953 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7"} Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.827965 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.828868 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.828904 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.828921 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.831091 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.831854 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.831904 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.831914 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:49 crc kubenswrapper[4935]: W1005 06:52:49.867349 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.59:6443: connect: connection refused Oct 05 06:52:49 crc kubenswrapper[4935]: E1005 06:52:49.867435 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.59:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.967102 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.969308 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.969367 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.969383 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:49 crc kubenswrapper[4935]: I1005 06:52:49.969413 4935 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:52:49 crc kubenswrapper[4935]: E1005 06:52:49.969957 4935 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.59:6443: connect: connection refused" node="crc" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.240328 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.818322 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.834994 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa"} Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.835076 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.835973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.836010 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.836024 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.837877 4935 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d45c1c61dd37f5b09210065acc7f2c3ebdbc9d8dc611088cca7dd98acb743e32" exitCode=0 Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.837974 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.838016 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.838035 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d45c1c61dd37f5b09210065acc7f2c3ebdbc9d8dc611088cca7dd98acb743e32"} Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.838072 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.838077 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.838987 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839044 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839326 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839367 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839380 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839693 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839718 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839727 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839732 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839747 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:50 crc kubenswrapper[4935]: I1005 06:52:50.839756 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.844336 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f4b359858a2f67cb5dbe8a7146e70ebe2e092bc63bfeb128d3beb7a6392aa2e9"} Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.844396 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f23abd0f18bf95480abb33abcccc3e0117a5be9d9bc2293cf483ee7f3f97ee52"} Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.844410 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ed93f0f101628de50c9a5063d2b08495f126fe076deb74fd6f7ac7b97fc4fe36"} Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.844420 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3c36896e50fae5df884c5a13be14a2ca689233184e8ba6fc76f95826ee8bfddf"} Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.844441 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.844462 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.844497 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.845452 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.845471 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.845489 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.845500 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.845502 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:51 crc kubenswrapper[4935]: I1005 06:52:51.845516 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.066078 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.066394 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.068167 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.068226 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.068243 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.796861 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.808041 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.856560 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3bb8189411d4b5999e6627434ea58ee203b7755b9797fcffd4ecf796717a587e"} Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.856590 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.856645 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.858260 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.858329 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.858330 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.858357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.858392 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.858501 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.862188 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.862375 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.862429 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.863566 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.863632 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:52 crc kubenswrapper[4935]: I1005 06:52:52.863656 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.171020 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.172429 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.172503 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.172519 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.172557 4935 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.241461 4935 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.241603 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.859498 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.859506 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.860949 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.861013 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.861033 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.861506 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.861558 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:53 crc kubenswrapper[4935]: I1005 06:52:53.861572 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.804619 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.862370 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.863704 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.863755 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.863774 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.973789 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.974115 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.975833 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.975984 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:54 crc kubenswrapper[4935]: I1005 06:52:54.976008 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.791488 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.830226 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.830352 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.830386 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.834183 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.834231 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.834246 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.865031 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.866071 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.866108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:55 crc kubenswrapper[4935]: I1005 06:52:55.866119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:52:56 crc kubenswrapper[4935]: E1005 06:52:56.863746 4935 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 05 06:52:56 crc kubenswrapper[4935]: I1005 06:52:56.974432 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:52:56 crc kubenswrapper[4935]: I1005 06:52:56.974582 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:52:56 crc kubenswrapper[4935]: I1005 06:52:56.975941 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:52:56 crc kubenswrapper[4935]: I1005 06:52:56.976037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:52:56 crc kubenswrapper[4935]: I1005 06:52:56.976060 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:00 crc kubenswrapper[4935]: W1005 06:53:00.547091 4935 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.547225 4935 trace.go:236] Trace[845368881]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 06:52:50.545) (total time: 10001ms): Oct 05 06:53:00 crc kubenswrapper[4935]: Trace[845368881]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:53:00.547) Oct 05 06:53:00 crc kubenswrapper[4935]: Trace[845368881]: [10.001266481s] [10.001266481s] END Oct 05 06:53:00 crc kubenswrapper[4935]: E1005 06:53:00.547260 4935 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.711431 4935 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.840859 4935 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:39642->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.841020 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:39642->192.168.126.11:17697: read: connection reset by peer" Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.882001 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.883571 4935 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa" exitCode=255 Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.883635 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa"} Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.883803 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.884973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.885020 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.885038 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:00 crc kubenswrapper[4935]: I1005 06:53:00.885984 4935 scope.go:117] "RemoveContainer" containerID="42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa" Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.155229 4935 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.155304 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.160634 4935 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.160685 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.896493 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.898844 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6"} Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.899031 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.900024 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.900081 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:01 crc kubenswrapper[4935]: I1005 06:53:01.900095 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:02 crc kubenswrapper[4935]: I1005 06:53:02.079971 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:53:02 crc kubenswrapper[4935]: I1005 06:53:02.080162 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:53:02 crc kubenswrapper[4935]: I1005 06:53:02.081742 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:02 crc kubenswrapper[4935]: I1005 06:53:02.081790 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:02 crc kubenswrapper[4935]: I1005 06:53:02.081808 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:03 crc kubenswrapper[4935]: I1005 06:53:03.240687 4935 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 05 06:53:03 crc kubenswrapper[4935]: I1005 06:53:03.240799 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 06:53:04 crc kubenswrapper[4935]: I1005 06:53:04.448779 4935 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.813519 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.813707 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.815207 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.815247 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.815264 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.825123 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.834591 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.834744 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.834785 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.835985 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.836052 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.836068 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.838830 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.909864 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.909881 4935 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.910985 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.911037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.911079 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.911090 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.911042 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:05 crc kubenswrapper[4935]: I1005 06:53:05.911133 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.138546 4935 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.144527 4935 trace.go:236] Trace[1978421475]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 06:52:53.463) (total time: 12680ms): Oct 05 06:53:06 crc kubenswrapper[4935]: Trace[1978421475]: ---"Objects listed" error: 12680ms (06:53:06.144) Oct 05 06:53:06 crc kubenswrapper[4935]: Trace[1978421475]: [12.680807886s] [12.680807886s] END Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.144573 4935 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.145385 4935 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.145604 4935 trace.go:236] Trace[910740179]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 06:52:54.672) (total time: 11473ms): Oct 05 06:53:06 crc kubenswrapper[4935]: Trace[910740179]: ---"Objects listed" error: 11473ms (06:53:06.145) Oct 05 06:53:06 crc kubenswrapper[4935]: Trace[910740179]: [11.473058661s] [11.473058661s] END Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.145650 4935 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.146718 4935 trace.go:236] Trace[513291463]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 06:52:51.361) (total time: 14785ms): Oct 05 06:53:06 crc kubenswrapper[4935]: Trace[513291463]: ---"Objects listed" error: 14784ms (06:53:06.146) Oct 05 06:53:06 crc kubenswrapper[4935]: Trace[513291463]: [14.785130189s] [14.785130189s] END Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.146750 4935 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.150395 4935 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.150721 4935 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.152101 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.152179 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.152197 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.152223 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.152240 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.168274 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.171609 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.171656 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.171669 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.171689 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.171702 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.184608 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.187836 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.187864 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.187877 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.187917 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.187938 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.198172 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.201881 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.201928 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.201939 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.201956 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.201965 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.213257 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.216630 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.216656 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.216665 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.216680 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.216689 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.240535 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.240643 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.242158 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.242186 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.242195 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.242212 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.242223 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.344409 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.344465 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.344479 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.344503 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.344518 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.447182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.447229 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.447239 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.447262 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.447273 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.550095 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.550167 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.550223 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.550250 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.550262 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.652757 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.652802 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.652817 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.652842 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.652854 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.703798 4935 apiserver.go:52] "Watching apiserver" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.708569 4935 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.708836 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-gth9x","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.709491 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-gth9x" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.709786 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.709988 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.709852 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.710698 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.710960 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.711017 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.711069 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.711587 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.711636 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.712101 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.712234 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.712414 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.712592 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.712766 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.713178 4935 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.714117 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.714496 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.714660 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.715016 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.715159 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.715296 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.715419 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.732251 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.743690 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749559 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749623 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749660 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749696 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749726 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749750 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749814 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749849 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749879 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749930 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.749959 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750021 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750048 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750078 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750105 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750133 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750162 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750197 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750233 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750260 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750290 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750320 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750350 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750374 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750406 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750435 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750460 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750492 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750522 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750546 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750575 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750605 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750638 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750664 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750696 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750730 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750760 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750789 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750853 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750881 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750926 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750952 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.750978 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.751005 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.751031 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.751039 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.751475 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.751620 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.751793 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752076 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752239 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.751062 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752415 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752482 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752515 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752538 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752543 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752841 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752880 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754008 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754125 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754209 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754288 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754369 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754446 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754522 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754598 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754677 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754758 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754993 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.755339 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.755425 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756322 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756386 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752725 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756665 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752868 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.752993 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.753300 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.753465 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.753517 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.753711 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.753966 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754045 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754572 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754735 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754359 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754618 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.754943 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.755198 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.755236 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.755233 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.755338 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.755505 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.755519 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756217 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756402 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756444 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756552 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756677 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756733 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756365 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756737 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756372 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756746 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.756760 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757089 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757204 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757227 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757309 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757344 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757360 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757386 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757410 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757435 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757454 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757454 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757477 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757492 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757506 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757526 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757549 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757570 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757591 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757608 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757624 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757649 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757671 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757689 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757708 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757728 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757746 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757765 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757854 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757875 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757923 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757952 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757977 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758064 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758097 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758116 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758150 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758172 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758197 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758214 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758235 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758259 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758278 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758294 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758312 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758330 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758353 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758375 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758393 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758423 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758443 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758445 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758469 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758656 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758678 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758696 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758716 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758737 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758756 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758774 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758795 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758813 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758834 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758855 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758876 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758919 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758937 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758956 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758976 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759003 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759027 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759046 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759064 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759081 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759093 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759171 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759181 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759199 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759208 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759102 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759790 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759817 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759840 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759859 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759881 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759929 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759957 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759984 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760014 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760041 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760070 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760094 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760114 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760132 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760150 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760167 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760187 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760205 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760224 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760242 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760259 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760276 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760295 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760314 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760330 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760350 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760374 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760394 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760413 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760432 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760470 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760511 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760530 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760551 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760572 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760593 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760610 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760628 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760646 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760665 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760683 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760701 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760725 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760747 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760768 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760798 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760818 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760843 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760866 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760901 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760923 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760944 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760965 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760989 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761007 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761029 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761049 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761069 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761093 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761115 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761153 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761180 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761206 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761229 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/551313e5-e4b6-4462-86c7-c6b7087fba96-hosts-file\") pod \"node-resolver-gth9x\" (UID: \"551313e5-e4b6-4462-86c7-c6b7087fba96\") " pod="openshift-dns/node-resolver-gth9x" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761255 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761278 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761300 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761322 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761343 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761364 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761388 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761409 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761429 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wlzv\" (UniqueName: \"kubernetes.io/projected/551313e5-e4b6-4462-86c7-c6b7087fba96-kube-api-access-6wlzv\") pod \"node-resolver-gth9x\" (UID: \"551313e5-e4b6-4462-86c7-c6b7087fba96\") " pod="openshift-dns/node-resolver-gth9x" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761456 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761486 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761508 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761570 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761583 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761595 4935 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761607 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761617 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761630 4935 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761641 4935 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761652 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761662 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761674 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761685 4935 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761697 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761706 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761717 4935 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761728 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761738 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761748 4935 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761758 4935 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761768 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761781 4935 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761790 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761801 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761811 4935 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761822 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761832 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761843 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761853 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761864 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761872 4935 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761883 4935 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762594 4935 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762610 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762621 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762633 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762645 4935 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762656 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762671 4935 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762686 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762700 4935 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762714 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762725 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762736 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762746 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762757 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757562 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757910 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.757929 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758086 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758772 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758838 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.758873 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.763948 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.764044 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759085 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759107 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759157 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759395 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759638 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.759909 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760690 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760699 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760743 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760759 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760780 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760781 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760793 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760803 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.760847 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761259 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761295 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761244 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761331 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761426 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761417 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761509 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761754 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.761973 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762093 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762230 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762393 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762490 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.762645 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.763081 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.763092 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.763218 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.763309 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.763456 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.763721 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.763793 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.765448 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.765655 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.765733 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.766239 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.766262 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.766333 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.766674 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.766953 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.767011 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.767393 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.767873 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.768044 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.768131 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.768171 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.768505 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.768507 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.768762 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.768820 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.769089 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.769324 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.769344 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.770319 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.770343 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.770490 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.770779 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.770834 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.770859 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.770879 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.770917 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.770949 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:53:07.270860339 +0000 UTC m=+21.153487059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.771067 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.771167 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.771256 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.771487 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.771716 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.771824 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.772049 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.772071 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.772309 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.773046 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.773301 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.773400 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.773343 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.773694 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.773514 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.774015 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.774058 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.774161 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.774501 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.774560 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.775094 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.775150 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.775242 4935 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.775445 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:07.275300608 +0000 UTC m=+21.157927078 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.775548 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.775654 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.776080 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.776270 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.776498 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.777443 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.777614 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.778142 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.778260 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.778316 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.778531 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.779176 4935 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.779295 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.779793 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.780396 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.780434 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.780450 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.780915 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.781386 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.781422 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.781756 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.782396 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.781854 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.781944 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.782245 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.782308 4935 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.782592 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:07.282559563 +0000 UTC m=+21.165186023 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.782772 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.782930 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.784344 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.781989 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.782324 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.782388 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.784735 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.783244 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.783285 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.783317 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.783751 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.783899 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.784015 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.785172 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.785294 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.785478 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.783056 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.786197 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.786249 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.786578 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.787593 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.787947 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.787963 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.788180 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.788410 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.789147 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.790056 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.793575 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.793782 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.794544 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.796556 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.797537 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.800934 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.801570 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.801594 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.801612 4935 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.801699 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:07.301675567 +0000 UTC m=+21.184302027 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.802633 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.804796 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.804839 4935 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:06 crc kubenswrapper[4935]: E1005 06:53:06.810124 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:07.30995392 +0000 UTC m=+21.192580380 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.814523 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.816197 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.827512 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.827737 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.827758 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.834656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.837824 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.841165 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.842638 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.843142 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.843909 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.846202 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.848230 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.851130 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.857039 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.857532 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.858512 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.858565 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.859869 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.860782 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.862216 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.862435 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.862469 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.862483 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.862503 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.862523 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863060 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863124 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wlzv\" (UniqueName: \"kubernetes.io/projected/551313e5-e4b6-4462-86c7-c6b7087fba96-kube-api-access-6wlzv\") pod \"node-resolver-gth9x\" (UID: \"551313e5-e4b6-4462-86c7-c6b7087fba96\") " pod="openshift-dns/node-resolver-gth9x" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863216 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863239 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/551313e5-e4b6-4462-86c7-c6b7087fba96-hosts-file\") pod \"node-resolver-gth9x\" (UID: \"551313e5-e4b6-4462-86c7-c6b7087fba96\") " pod="openshift-dns/node-resolver-gth9x" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863288 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863289 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863302 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863338 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863348 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863357 4935 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863369 4935 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863378 4935 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863388 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863398 4935 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863409 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863403 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863417 4935 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863448 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863459 4935 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863124 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863471 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/551313e5-e4b6-4462-86c7-c6b7087fba96-hosts-file\") pod \"node-resolver-gth9x\" (UID: \"551313e5-e4b6-4462-86c7-c6b7087fba96\") " pod="openshift-dns/node-resolver-gth9x" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863502 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863531 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863547 4935 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863562 4935 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863575 4935 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863589 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863605 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863619 4935 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863631 4935 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863644 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863657 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863670 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863682 4935 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863693 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863705 4935 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863718 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863736 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863738 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863750 4935 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863789 4935 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863801 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.863812 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864035 4935 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864048 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864061 4935 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864074 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864087 4935 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864100 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864115 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864127 4935 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864142 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864327 4935 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864352 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864385 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864397 4935 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864407 4935 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864405 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864418 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864435 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864476 4935 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864709 4935 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864727 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864738 4935 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864747 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864774 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864784 4935 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864794 4935 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864805 4935 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864813 4935 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864822 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864831 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864841 4935 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864850 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864860 4935 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864872 4935 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864881 4935 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864908 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864921 4935 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864930 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864939 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864948 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864957 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864966 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864976 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864986 4935 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.864997 4935 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865006 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865016 4935 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865026 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865035 4935 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865044 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865054 4935 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865066 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865075 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865086 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865096 4935 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865105 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865114 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865124 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865134 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865143 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865152 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865162 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865171 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865180 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865188 4935 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865198 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865207 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865216 4935 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865226 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865236 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865250 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865259 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865271 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865279 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865289 4935 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865299 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865312 4935 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865321 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865330 4935 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865341 4935 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865350 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865362 4935 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865372 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865381 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865393 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865403 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865413 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865472 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865507 4935 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865527 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865546 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865563 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865577 4935 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865591 4935 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865601 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865603 4935 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865650 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865663 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865677 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865691 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865734 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865748 4935 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865761 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865775 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865815 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865830 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865845 4935 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865880 4935 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865914 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865928 4935 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.865966 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.866017 4935 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.866061 4935 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.866077 4935 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.866092 4935 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.866127 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.866142 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.866158 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.866180 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.867225 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.867652 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.870332 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.871019 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.871548 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.877059 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.880350 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.881481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wlzv\" (UniqueName: \"kubernetes.io/projected/551313e5-e4b6-4462-86c7-c6b7087fba96-kube-api-access-6wlzv\") pod \"node-resolver-gth9x\" (UID: \"551313e5-e4b6-4462-86c7-c6b7087fba96\") " pod="openshift-dns/node-resolver-gth9x" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.881999 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.882515 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.882689 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.883948 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.885091 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.886055 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.886837 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.887748 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.888332 4935 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.888508 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.890736 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.891417 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.892094 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.894845 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.896140 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.897546 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.898082 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.899550 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.900561 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.901560 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.902303 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.903414 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.904657 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.906347 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.907041 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.907152 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.908291 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.909318 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.910185 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.911435 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.912038 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.913084 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.914591 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.916023 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.919895 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.932615 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.956316 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.965017 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.965070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.965082 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.965101 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.965117 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:06Z","lastTransitionTime":"2025-10-05T06:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.967088 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.968862 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.992953 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:06 crc kubenswrapper[4935]: I1005 06:53:06.995440 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.022733 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-gth9x" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.029131 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-j8cwh"] Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.029555 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.029746 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.032448 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.032576 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.032841 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.037205 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.039683 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.040362 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 05 06:53:07 crc kubenswrapper[4935]: W1005 06:53:07.044150 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod551313e5_e4b6_4462_86c7_c6b7087fba96.slice/crio-1d0344df9191b8642e953d1d53503f40ae4b862dbdc9e6863e9a1d17fbef5185 WatchSource:0}: Error finding container 1d0344df9191b8642e953d1d53503f40ae4b862dbdc9e6863e9a1d17fbef5185: Status 404 returned error can't find the container with id 1d0344df9191b8642e953d1d53503f40ae4b862dbdc9e6863e9a1d17fbef5185 Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.044683 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.045793 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-687xm"] Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.049804 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.063078 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.063476 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-cz2rj"] Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.078238 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.078309 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.078329 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.078329 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.079077 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ns4c5"] Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.079176 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.079493 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.081419 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.088273 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.088391 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.088278 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.088545 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.088669 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.088706 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.088836 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.088976 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.089097 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.089531 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.089559 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.089570 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.089588 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.089599 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.110920 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.130625 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.148124 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.162759 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.172998 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179412 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-daemon-config\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179459 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179492 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-env-overrides\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179512 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-script-lib\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179535 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-netns\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179563 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62823424-a6c9-4733-b50a-6f3ecb4a55a2-mcd-auth-proxy-config\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179583 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/62823424-a6c9-4733-b50a-6f3ecb4a55a2-rootfs\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179603 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-log-socket\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179633 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56m5n\" (UniqueName: \"kubernetes.io/projected/2b9ff491-4f10-4f1b-822b-a8c46871ce60-kube-api-access-56m5n\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.179680 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-ovn\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180644 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-k8s-cni-cncf-io\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180725 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-etc-kubernetes\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180750 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-slash\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180807 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-cni-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180834 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-os-release\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180855 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-systemd-units\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180923 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-multus-certs\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180948 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cnibin\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.180981 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-var-lib-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181002 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-system-cni-dir\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181027 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62823424-a6c9-4733-b50a-6f3ecb4a55a2-proxy-tls\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181050 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-systemd\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181071 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-netd\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181093 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovn-node-metrics-cert\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181114 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b9ff491-4f10-4f1b-822b-a8c46871ce60-cni-binary-copy\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181136 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181158 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-kubelet\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181187 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-bin\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181213 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-config\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181136 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181287 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-os-release\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181316 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-conf-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181336 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-ovn-kubernetes\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181363 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7s48\" (UniqueName: \"kubernetes.io/projected/62823424-a6c9-4733-b50a-6f3ecb4a55a2-kube-api-access-b7s48\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181382 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-system-cni-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181397 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-socket-dir-parent\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181419 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-cni-bin\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181438 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-cni-multus\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181455 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-hostroot\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181481 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cni-binary-copy\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181496 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-node-log\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181538 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rlhk\" (UniqueName: \"kubernetes.io/projected/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-kube-api-access-2rlhk\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181558 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-netns\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181574 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181615 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzplz\" (UniqueName: \"kubernetes.io/projected/019e7dde-1935-4b4c-8312-99f52b6d22fc-kube-api-access-nzplz\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181631 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-cnibin\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181654 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-kubelet\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181670 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.181686 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-etc-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.192849 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.192933 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.192948 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.192965 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.192975 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.194838 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.205499 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.215362 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.227254 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.241651 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.259777 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.269534 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.282692 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.282871 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.282944 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:53:08.282880606 +0000 UTC m=+22.165507076 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.282995 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283027 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-env-overrides\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283098 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-script-lib\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283125 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-daemon-config\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283166 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-netns\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283193 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62823424-a6c9-4733-b50a-6f3ecb4a55a2-mcd-auth-proxy-config\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283236 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/62823424-a6c9-4733-b50a-6f3ecb4a55a2-rootfs\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283260 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-log-socket\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283297 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56m5n\" (UniqueName: \"kubernetes.io/projected/2b9ff491-4f10-4f1b-822b-a8c46871ce60-kube-api-access-56m5n\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283319 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-ovn\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283342 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-k8s-cni-cncf-io\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283366 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-etc-kubernetes\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283390 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-slash\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283413 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-cni-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283438 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-os-release\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283468 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-multus-certs\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283494 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cnibin\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283557 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-systemd-units\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283560 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-ovn\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283629 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-var-lib-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283668 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-system-cni-dir\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283694 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62823424-a6c9-4733-b50a-6f3ecb4a55a2-proxy-tls\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283697 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-etc-kubernetes\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283717 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-systemd\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283773 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-netd\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283798 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovn-node-metrics-cert\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283830 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283861 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-kubelet\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283857 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-env-overrides\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283887 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-bin\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283744 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-systemd\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283931 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b9ff491-4f10-4f1b-822b-a8c46871ce60-cni-binary-copy\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284003 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-os-release\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284023 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-conf-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284060 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-ovn-kubernetes\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284076 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-config\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284096 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-system-cni-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284132 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-socket-dir-parent\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284151 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-cni-bin\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284171 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-cni-multus\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284203 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-hostroot\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284223 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cni-binary-copy\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284240 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7s48\" (UniqueName: \"kubernetes.io/projected/62823424-a6c9-4733-b50a-6f3ecb4a55a2-kube-api-access-b7s48\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284257 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-node-log\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284294 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rlhk\" (UniqueName: \"kubernetes.io/projected/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-kube-api-access-2rlhk\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284316 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-netns\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284338 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284410 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzplz\" (UniqueName: \"kubernetes.io/projected/019e7dde-1935-4b4c-8312-99f52b6d22fc-kube-api-access-nzplz\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284455 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284475 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-kubelet\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284526 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284543 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-etc-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284557 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-cnibin\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284572 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284611 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-daemon-config\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b9ff491-4f10-4f1b-822b-a8c46871ce60-cni-binary-copy\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.284657 4935 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284713 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-kubelet\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284781 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-bin\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.284770 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:08.284711626 +0000 UTC m=+22.167338086 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284833 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-cni-multus\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284852 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-hostroot\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.284879 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-slash\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285068 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-cni-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285103 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-netd\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285170 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cnibin\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285180 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-ovn-kubernetes\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285215 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-multus-certs\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285193 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-system-cni-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285232 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-socket-dir-parent\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285335 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-os-release\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285368 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-multus-conf-dir\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285366 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-os-release\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285404 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285438 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-cni-bin\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285544 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-node-log\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285620 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-cnibin\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285626 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-etc-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.285656 4935 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285721 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.285728 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:08.285705532 +0000 UTC m=+22.168331992 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285756 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-var-lib-kubelet\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285664 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-netns\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283670 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-k8s-cni-cncf-io\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285787 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-system-cni-dir\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285832 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-var-lib-openvswitch\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.283637 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-systemd-units\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285913 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/62823424-a6c9-4733-b50a-6f3ecb4a55a2-rootfs\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285949 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-log-socket\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.285999 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b9ff491-4f10-4f1b-822b-a8c46871ce60-host-run-netns\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.286028 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.286537 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-cni-binary-copy\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.286747 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-script-lib\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.286980 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-config\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.287993 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62823424-a6c9-4733-b50a-6f3ecb4a55a2-mcd-auth-proxy-config\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.288918 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.290265 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62823424-a6c9-4733-b50a-6f3ecb4a55a2-proxy-tls\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.291551 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovn-node-metrics-cert\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.299475 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.299513 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.299527 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.299547 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.299558 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.304803 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rlhk\" (UniqueName: \"kubernetes.io/projected/8e242b9a-8af9-4ddc-9c51-27d1432c80fd-kube-api-access-2rlhk\") pod \"multus-additional-cni-plugins-cz2rj\" (UID: \"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\") " pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.305603 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzplz\" (UniqueName: \"kubernetes.io/projected/019e7dde-1935-4b4c-8312-99f52b6d22fc-kube-api-access-nzplz\") pod \"ovnkube-node-ns4c5\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.306069 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7s48\" (UniqueName: \"kubernetes.io/projected/62823424-a6c9-4733-b50a-6f3ecb4a55a2-kube-api-access-b7s48\") pod \"machine-config-daemon-j8cwh\" (UID: \"62823424-a6c9-4733-b50a-6f3ecb4a55a2\") " pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.308319 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56m5n\" (UniqueName: \"kubernetes.io/projected/2b9ff491-4f10-4f1b-822b-a8c46871ce60-kube-api-access-56m5n\") pod \"multus-687xm\" (UID: \"2b9ff491-4f10-4f1b-822b-a8c46871ce60\") " pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.308625 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.319823 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.338383 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.343599 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.350141 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.362643 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.363436 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-687xm" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.379409 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.385507 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.385627 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.385733 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.385765 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.385779 4935 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.385834 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:08.385814214 +0000 UTC m=+22.268440674 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.386008 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.386039 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.386054 4935 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:07 crc kubenswrapper[4935]: E1005 06:53:07.386114 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:08.386095062 +0000 UTC m=+22.268721702 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.397399 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.402124 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.402161 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.402170 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.402191 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.402203 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.411021 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.434140 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.445914 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:07 crc kubenswrapper[4935]: W1005 06:53:07.471022 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod019e7dde_1935_4b4c_8312_99f52b6d22fc.slice/crio-02438eba25ad62d4e9a84e0e6b032ee5db5a92a64239d6a9e862635d9cbb9442 WatchSource:0}: Error finding container 02438eba25ad62d4e9a84e0e6b032ee5db5a92a64239d6a9e862635d9cbb9442: Status 404 returned error can't find the container with id 02438eba25ad62d4e9a84e0e6b032ee5db5a92a64239d6a9e862635d9cbb9442 Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.504916 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.504973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.504983 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.505003 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.505014 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.607830 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.607875 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.607913 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.607936 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.607952 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.710669 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.710715 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.710725 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.710742 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.710754 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.813839 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.813914 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.813929 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.813962 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.813978 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.915783 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.915818 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.915828 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.915847 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.915862 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:07Z","lastTransitionTime":"2025-10-05T06:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.917442 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.917490 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9c3f76ccdbe5b7cf41a1be18c74f20393bc055a8484ce9ea5d37fc060a834713"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.919570 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa" exitCode=0 Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.919619 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.919649 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"02438eba25ad62d4e9a84e0e6b032ee5db5a92a64239d6a9e862635d9cbb9442"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.921460 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-687xm" event={"ID":"2b9ff491-4f10-4f1b-822b-a8c46871ce60","Type":"ContainerStarted","Data":"7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.921520 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-687xm" event={"ID":"2b9ff491-4f10-4f1b-822b-a8c46871ce60","Type":"ContainerStarted","Data":"85f757fd5d9e99d9ef2172bd7ba18603ee41542a1de772000230bca973164528"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.922437 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"eb86dd2fc12ddda096ab46df7d1e5c2746fc37589a4afb59c734c101b6da0fef"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.923805 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-gth9x" event={"ID":"551313e5-e4b6-4462-86c7-c6b7087fba96","Type":"ContainerStarted","Data":"858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.923834 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-gth9x" event={"ID":"551313e5-e4b6-4462-86c7-c6b7087fba96","Type":"ContainerStarted","Data":"1d0344df9191b8642e953d1d53503f40ae4b862dbdc9e6863e9a1d17fbef5185"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.929868 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerStarted","Data":"7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.929965 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerStarted","Data":"63e9561b15fbce33baa2de4f4053036b08f9484866ef2d84cc3a074af43d9206"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.931902 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.931939 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.931954 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"255f12459f7fbf27becf85dbd015ea110ff595cfb86c9b3f140e294dcef5e44b"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.934407 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.934482 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.934495 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cc23644b975fb92e89e6a61d76fc86a4ed19d122284a7145e20a198d720146c7"} Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.934764 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.945265 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.954080 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.973164 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:07 crc kubenswrapper[4935]: I1005 06:53:07.990410 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.005005 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.018595 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.018656 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.018672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.018692 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.018706 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.022035 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.047847 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.066157 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.085318 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.119451 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.121380 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.121426 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.121502 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.121521 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.121534 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.139025 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.150356 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.164812 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.178925 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.198405 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.218647 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.223748 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.223853 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.223935 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.224036 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.224132 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.241715 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.255935 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.270871 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.283066 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.293653 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.293789 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:53:10.293767667 +0000 UTC m=+24.176394127 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.294007 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.294149 4935 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.294208 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:10.294197258 +0000 UTC m=+24.176823718 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.294270 4935 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.294396 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:10.294365583 +0000 UTC m=+24.176992103 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.294754 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.297549 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.325930 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.342802 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.342859 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.342875 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.342913 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.342930 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.366037 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.395503 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.395568 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.395742 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.395793 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.395755 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.395848 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.395864 4935 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.395809 4935 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.395955 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:10.395932574 +0000 UTC m=+24.278559204 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.396021 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:10.395989595 +0000 UTC m=+24.278616245 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.446801 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.446851 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.446864 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.446906 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.446921 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.549605 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.549995 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.550066 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.550131 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.550196 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.653589 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.653666 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.653686 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.653717 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.653737 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.758336 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.758412 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.758433 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.758462 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.758486 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.776629 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.776667 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.776652 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.776834 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.777023 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:08 crc kubenswrapper[4935]: E1005 06:53:08.777270 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.864334 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.864403 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.864422 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.864449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.864469 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.940107 4935 generic.go:334] "Generic (PLEG): container finished" podID="8e242b9a-8af9-4ddc-9c51-27d1432c80fd" containerID="7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a" exitCode=0 Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.940202 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerDied","Data":"7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.944184 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.944236 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.964627 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.967322 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.967379 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.967398 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.967425 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.967445 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:08Z","lastTransitionTime":"2025-10-05T06:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:08 crc kubenswrapper[4935]: I1005 06:53:08.986520 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.010387 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.030137 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.057743 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.070917 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.070973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.070985 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.071006 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.071019 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.074483 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.087069 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.100528 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.115075 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.127665 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.147460 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.163340 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.173991 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.174038 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.174051 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.174070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.174085 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.276991 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.277034 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.277048 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.277069 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.277083 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.333426 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-w5h5q"] Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.333904 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.335706 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.336564 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.338170 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.353011 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.363937 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.364728 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.379633 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.379700 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.379722 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.379746 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.379766 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.380186 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.394544 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.406220 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.407805 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fc6754de-5445-4fe9-8707-2bca9b3e955a-serviceca\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.407885 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc6754de-5445-4fe9-8707-2bca9b3e955a-host\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.407931 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcsgf\" (UniqueName: \"kubernetes.io/projected/fc6754de-5445-4fe9-8707-2bca9b3e955a-kube-api-access-lcsgf\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.420528 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.432269 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.446070 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.465666 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.483054 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.483108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.483121 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.483145 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.483161 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.484711 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.496250 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.509199 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fc6754de-5445-4fe9-8707-2bca9b3e955a-serviceca\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.509261 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc6754de-5445-4fe9-8707-2bca9b3e955a-host\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.509283 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcsgf\" (UniqueName: \"kubernetes.io/projected/fc6754de-5445-4fe9-8707-2bca9b3e955a-kube-api-access-lcsgf\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.509252 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.509682 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc6754de-5445-4fe9-8707-2bca9b3e955a-host\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.511778 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fc6754de-5445-4fe9-8707-2bca9b3e955a-serviceca\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.530568 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcsgf\" (UniqueName: \"kubernetes.io/projected/fc6754de-5445-4fe9-8707-2bca9b3e955a-kube-api-access-lcsgf\") pod \"node-ca-w5h5q\" (UID: \"fc6754de-5445-4fe9-8707-2bca9b3e955a\") " pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.541480 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.586043 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.586109 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.586129 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.586155 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.586174 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.648793 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-w5h5q" Oct 05 06:53:09 crc kubenswrapper[4935]: W1005 06:53:09.662338 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc6754de_5445_4fe9_8707_2bca9b3e955a.slice/crio-f62042d9246585bffb177850aea84e60c3843d1d993806719689538ead212cc2 WatchSource:0}: Error finding container f62042d9246585bffb177850aea84e60c3843d1d993806719689538ead212cc2: Status 404 returned error can't find the container with id f62042d9246585bffb177850aea84e60c3843d1d993806719689538ead212cc2 Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.689504 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.689537 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.689548 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.689567 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.689579 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.792963 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.793249 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.793270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.793289 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.793301 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.896244 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.896280 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.896287 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.896303 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.896312 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.947370 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-w5h5q" event={"ID":"fc6754de-5445-4fe9-8707-2bca9b3e955a","Type":"ContainerStarted","Data":"f62042d9246585bffb177850aea84e60c3843d1d993806719689538ead212cc2"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.949334 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerStarted","Data":"34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.951981 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.952003 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.952012 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e"} Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.964379 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.976748 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.993182 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:09Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.998370 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.998435 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.998449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.998487 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:09 crc kubenswrapper[4935]: I1005 06:53:09.998501 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:09Z","lastTransitionTime":"2025-10-05T06:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.019276 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.032916 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.043346 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.057591 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.074189 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.090257 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.100974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.101029 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.101040 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.101058 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.101072 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.105337 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.117560 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.129713 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.140490 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.203698 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.203743 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.203756 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.203777 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.203789 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.245123 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.249205 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.254861 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.261916 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.275946 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.286846 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.300638 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.306242 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.306288 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.306301 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.306344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.306357 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.319454 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.319719 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.319858 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.319902 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:53:14.319855316 +0000 UTC m=+28.202481786 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.319941 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.320028 4935 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.320053 4935 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.320110 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:14.320086792 +0000 UTC m=+28.202713412 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.320145 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:14.320124633 +0000 UTC m=+28.202751303 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.334183 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.346528 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.361927 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.378216 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.392145 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.408435 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.408493 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.408506 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.408535 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.408552 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.409556 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.420572 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.420630 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.420853 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.420879 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.420908 4935 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.420978 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:14.420958264 +0000 UTC m=+28.303584724 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.421184 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.421273 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.421350 4935 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.421454 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:14.421441587 +0000 UTC m=+28.304068047 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.437758 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.461997 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.485254 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.498613 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.509774 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.511247 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.511288 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.511300 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.511320 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.511333 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.523517 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.546536 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.562408 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.578691 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.614179 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.614215 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.614228 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.614246 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.614259 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.615861 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.655123 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.697765 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.717061 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.717109 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.717121 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.717140 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.717153 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.736732 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.776831 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.776872 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.776991 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.777067 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.777174 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.777322 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:10 crc kubenswrapper[4935]: E1005 06:53:10.777405 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.816352 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.820017 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.820052 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.820062 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.820079 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.820091 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.855018 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.922839 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.922881 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.922915 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.922938 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.922953 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:10Z","lastTransitionTime":"2025-10-05T06:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.956681 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-w5h5q" event={"ID":"fc6754de-5445-4fe9-8707-2bca9b3e955a","Type":"ContainerStarted","Data":"b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.958952 4935 generic.go:334] "Generic (PLEG): container finished" podID="8e242b9a-8af9-4ddc-9c51-27d1432c80fd" containerID="34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13" exitCode=0 Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.959024 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerDied","Data":"34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.964914 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.966958 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a"} Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.975527 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:10 crc kubenswrapper[4935]: I1005 06:53:10.987553 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:10Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.004911 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.022919 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.025647 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.025672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.025680 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.025695 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.025704 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.055555 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.093607 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.128903 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.128937 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.128945 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.128963 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.128974 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.133190 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.178146 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.219465 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.231688 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.231725 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.231736 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.231754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.231766 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.255540 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.292882 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.333708 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.333954 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.334070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.334155 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.334223 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.339206 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.385165 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.418489 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.437034 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.437080 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.437089 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.437108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.437120 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.456832 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.496967 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.534856 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.540158 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.540360 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.540487 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.540626 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.540763 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.574300 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.622324 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.643094 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.643128 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.643141 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.643159 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.643170 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.655331 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.697164 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.737260 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.746265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.746310 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.746323 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.746344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.746361 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.773936 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.816332 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.848510 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.848560 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.848571 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.848670 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.848684 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.864328 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.897291 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.933968 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.951548 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.951590 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.951601 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.951619 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.951631 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:11Z","lastTransitionTime":"2025-10-05T06:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.974115 4935 generic.go:334] "Generic (PLEG): container finished" podID="8e242b9a-8af9-4ddc-9c51-27d1432c80fd" containerID="61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb" exitCode=0 Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.974173 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerDied","Data":"61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb"} Oct 05 06:53:11 crc kubenswrapper[4935]: I1005 06:53:11.983343 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:11Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.016528 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.054937 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.054999 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.055017 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.055039 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.055055 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.057817 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.093656 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.137451 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.157754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.157792 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.157802 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.157818 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.157830 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.180247 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.219698 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.254961 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.261101 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.261144 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.261155 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.261173 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.261185 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.296239 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.336787 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.363450 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.363512 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.363527 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.363548 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.363562 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.381066 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.420345 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.452349 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.466721 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.466882 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.466917 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.466932 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.466943 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.499271 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.534548 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.569306 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.569344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.569357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.569375 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.569386 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.671715 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.671776 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.671793 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.671812 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.671826 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.774412 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.774449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.774459 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.774474 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.774488 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.776973 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.777065 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.777164 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:12 crc kubenswrapper[4935]: E1005 06:53:12.777320 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:12 crc kubenswrapper[4935]: E1005 06:53:12.777509 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:12 crc kubenswrapper[4935]: E1005 06:53:12.777629 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.877452 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.877522 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.877542 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.877570 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.877589 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.979930 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.979997 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.980014 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.980044 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.980065 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:12Z","lastTransitionTime":"2025-10-05T06:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.984392 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5"} Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.987761 4935 generic.go:334] "Generic (PLEG): container finished" podID="8e242b9a-8af9-4ddc-9c51-27d1432c80fd" containerID="aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771" exitCode=0 Oct 05 06:53:12 crc kubenswrapper[4935]: I1005 06:53:12.987830 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerDied","Data":"aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.009156 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.025866 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.046304 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.060770 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.076304 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.082760 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.082802 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.082812 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.082829 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.082842 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.099102 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.123985 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.147328 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.172677 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.185801 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.185863 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.185878 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.185929 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.185953 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.189261 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.212290 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.226571 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.239327 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.254514 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.288310 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.288344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.288353 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.288367 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.288379 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.391580 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.391653 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.391663 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.391706 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.391723 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.494282 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.494327 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.494336 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.494354 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.494365 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.597592 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.597663 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.597678 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.597707 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.597725 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.700673 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.700748 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.700761 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.700783 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.700800 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.804216 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.804275 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.804294 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.804320 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.804341 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.907390 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.907468 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.907510 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.907555 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.907580 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:13Z","lastTransitionTime":"2025-10-05T06:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.996200 4935 generic.go:334] "Generic (PLEG): container finished" podID="8e242b9a-8af9-4ddc-9c51-27d1432c80fd" containerID="9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62" exitCode=0 Oct 05 06:53:13 crc kubenswrapper[4935]: I1005 06:53:13.996252 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerDied","Data":"9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.010366 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.010426 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.010446 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.010472 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.010492 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.024367 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.046826 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.069595 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.088546 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.104828 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.113037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.113116 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.113142 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.113172 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.113194 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.126397 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.141548 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.153115 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.168112 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.180729 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.196953 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.213122 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.215918 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.215957 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.215967 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.215983 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.215994 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.225324 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.238034 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:14Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.318446 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.318491 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.318502 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.318516 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.318526 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.363922 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.364082 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:53:22.364055948 +0000 UTC m=+36.246682408 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.364122 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.364151 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.364235 4935 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.364291 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:22.364279714 +0000 UTC m=+36.246906174 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.364802 4935 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.364853 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:22.364843839 +0000 UTC m=+36.247470299 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.421841 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.421910 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.421921 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.421939 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.421949 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.465306 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.465350 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.465502 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.465520 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.465531 4935 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.465573 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.465624 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.465637 4935 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.465591 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:22.465570467 +0000 UTC m=+36.348196927 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.465746 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:22.465722101 +0000 UTC m=+36.348348551 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.525163 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.525419 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.525428 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.525444 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.525454 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.628960 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.629038 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.629063 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.629093 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.629119 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.732596 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.732646 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.732678 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.732697 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.732711 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.776574 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.776709 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.776756 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.776812 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.776929 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:14 crc kubenswrapper[4935]: E1005 06:53:14.777023 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.834959 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.834996 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.835007 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.835024 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.835037 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.937951 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.938009 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.938022 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.938050 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:14 crc kubenswrapper[4935]: I1005 06:53:14.938067 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:14Z","lastTransitionTime":"2025-10-05T06:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.006191 4935 generic.go:334] "Generic (PLEG): container finished" podID="8e242b9a-8af9-4ddc-9c51-27d1432c80fd" containerID="b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a" exitCode=0 Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.006268 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerDied","Data":"b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.012054 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.012654 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.028987 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.041094 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.041146 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.041159 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.041184 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.041197 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.060373 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.064976 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.076162 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.090714 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.101869 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.115818 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.128644 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.142456 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.144085 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.144185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.144248 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.144311 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.144382 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.157489 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.188237 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.211248 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.225394 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.247529 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.247570 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.247580 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.247597 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.247607 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.247833 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.261080 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.272879 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.287666 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.302266 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.314774 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.327767 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.341746 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.350462 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.350499 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.350508 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.350525 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.350535 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.367380 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.392101 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.416275 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.430772 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.442753 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.453286 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.453327 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.453340 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.453356 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.453368 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.453862 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.463787 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.474132 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:15Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.556878 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.557464 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.557599 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.557800 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.557976 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.661170 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.661210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.661219 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.661234 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.661244 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.762982 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.763033 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.763043 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.763060 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.763256 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.865550 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.865620 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.865642 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.865666 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.865683 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.968368 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.968437 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.968462 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.968496 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:15 crc kubenswrapper[4935]: I1005 06:53:15.968521 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:15Z","lastTransitionTime":"2025-10-05T06:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.025654 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" event={"ID":"8e242b9a-8af9-4ddc-9c51-27d1432c80fd","Type":"ContainerStarted","Data":"c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.025998 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.026855 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.049244 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.055297 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.064553 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.078351 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.078403 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.078415 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.078439 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.078454 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.080169 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.092597 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.111709 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.143181 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.161465 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.172717 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.181083 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.181149 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.181161 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.181180 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.181190 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.191737 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.212321 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.230125 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.243677 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.257409 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.272884 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.284252 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.284293 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.284305 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.284322 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.284335 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.285774 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.303110 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.316006 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.327184 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.340173 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.352495 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.364350 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.379063 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.387300 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.387356 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.387370 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.387389 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.387403 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.397369 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.415340 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.429171 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.448605 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.463721 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.474542 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.490032 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.490100 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.490115 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.490139 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.490163 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.550691 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.550760 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.550781 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.550808 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.550827 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.572246 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.576989 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.577051 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.577070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.577098 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.577117 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.598961 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.603737 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.603775 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.603787 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.603809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.603821 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.629742 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.635405 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.635440 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.635449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.635464 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.635475 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.651961 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.656210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.656242 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.656257 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.656275 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.656287 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.668777 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.668995 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.671358 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.671427 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.671449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.671477 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.671498 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.775773 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.775819 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.775827 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.775843 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.775856 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.776186 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.776257 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.776310 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.776405 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.776517 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:16 crc kubenswrapper[4935]: E1005 06:53:16.776639 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.798226 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.816114 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.833313 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.847230 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.862396 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.878340 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.878405 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.878423 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.878449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.878467 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.883447 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.899958 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.920223 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.950857 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.971075 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.980600 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.981809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.981939 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.981960 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.981986 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.982004 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:16Z","lastTransitionTime":"2025-10-05T06:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:16 crc kubenswrapper[4935]: I1005 06:53:16.988758 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.000883 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.016931 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.029638 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.044710 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.060600 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.075648 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.085184 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.085223 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.085233 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.085250 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.085263 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.090721 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.111911 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.126597 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.136741 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.152301 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.163753 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.174933 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.184240 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.188053 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.188107 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.188120 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.188135 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.188146 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.197906 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.208633 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.228984 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.240581 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.290220 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.290269 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.290280 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.290296 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.290306 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.393441 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.393523 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.393555 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.393593 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.393636 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.497100 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.497356 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.497375 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.497402 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.497421 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.599973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.600016 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.600028 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.600045 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.600059 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.702149 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.702201 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.702213 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.702253 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.702266 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.804481 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.804716 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.804798 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.804870 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.804986 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.907790 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.907838 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.907851 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.907869 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:17 crc kubenswrapper[4935]: I1005 06:53:17.907901 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:17Z","lastTransitionTime":"2025-10-05T06:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.010323 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.010408 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.010423 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.010444 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.010458 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.032667 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.113295 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.113336 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.113347 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.113365 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.113376 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.216066 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.216111 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.216129 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.216149 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.216163 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.318963 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.319270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.319281 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.319301 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.319315 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.423166 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.423244 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.423262 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.423303 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.423323 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.526937 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.527014 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.527039 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.527069 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.527091 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.630702 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.630768 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.630779 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.630801 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.630812 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.733702 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.733752 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.733766 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.733788 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.733804 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.776619 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.776634 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:18 crc kubenswrapper[4935]: E1005 06:53:18.776868 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.776966 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:18 crc kubenswrapper[4935]: E1005 06:53:18.777116 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:18 crc kubenswrapper[4935]: E1005 06:53:18.777377 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.836209 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.836270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.836285 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.836310 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.836322 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.938972 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.939034 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.939043 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.939059 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:18 crc kubenswrapper[4935]: I1005 06:53:18.939070 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:18Z","lastTransitionTime":"2025-10-05T06:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.038608 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/0.log" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.041181 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.041223 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.041234 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.041250 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.041263 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.042182 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7" exitCode=1 Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.042245 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.043004 4935 scope.go:117] "RemoveContainer" containerID="c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.064117 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.082720 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.096535 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.111226 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.127326 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.139597 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.144585 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.144631 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.144643 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.144664 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.144677 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.155438 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.165052 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.177347 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.195560 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:18Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:53:18.101739 6206 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:53:18.101789 6206 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:53:18.101857 6206 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:53:18.101903 6206 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:53:18.101944 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:53:18.101972 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:53:18.101997 6206 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:53:18.101864 6206 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:53:18.101999 6206 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:53:18.102021 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:53:18.102032 6206 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:53:18.102062 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:53:18.102108 6206 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:53:18.102175 6206 factory.go:656] Stopping watch factory\\\\nI1005 06:53:18.102209 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:53:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.211546 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.223733 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.235547 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.247860 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.248046 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.248091 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.248108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.248133 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.248151 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.351033 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.351235 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.351269 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.351298 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.351317 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.442373 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd"] Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.443214 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.445052 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.446129 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.455517 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.455566 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.455582 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.455603 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.455619 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.470668 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.493228 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.512921 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.518846 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.518919 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.519168 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wqq5\" (UniqueName: \"kubernetes.io/projected/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-kube-api-access-5wqq5\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.519281 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.535283 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.554113 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.558469 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.558514 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.558531 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.558555 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.558573 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.573107 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.599010 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.618182 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.620996 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wqq5\" (UniqueName: \"kubernetes.io/projected/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-kube-api-access-5wqq5\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.621104 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.621146 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.621184 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.622310 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.622459 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.630945 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.644349 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.652666 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wqq5\" (UniqueName: \"kubernetes.io/projected/cb1942fa-f3fe-4aa9-99e3-09e12525ab6c-kube-api-access-5wqq5\") pod \"ovnkube-control-plane-749d76644c-znjdd\" (UID: \"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.661386 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.661420 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.661429 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.661443 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.661454 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.667176 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.687143 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.704635 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.725107 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.744775 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:18Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:53:18.101739 6206 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:53:18.101789 6206 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:53:18.101857 6206 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:53:18.101903 6206 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:53:18.101944 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:53:18.101972 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:53:18.101997 6206 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:53:18.101864 6206 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:53:18.101999 6206 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:53:18.102021 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:53:18.102032 6206 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:53:18.102062 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:53:18.102108 6206 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:53:18.102175 6206 factory.go:656] Stopping watch factory\\\\nI1005 06:53:18.102209 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:53:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.764731 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.764886 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.764967 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.764989 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.765016 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.765035 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.768293 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" Oct 05 06:53:19 crc kubenswrapper[4935]: W1005 06:53:19.790434 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb1942fa_f3fe_4aa9_99e3_09e12525ab6c.slice/crio-343e5005cdeb0ff6766f027efa57c372a04cd267a5631fbd060c9778fa310cbf WatchSource:0}: Error finding container 343e5005cdeb0ff6766f027efa57c372a04cd267a5631fbd060c9778fa310cbf: Status 404 returned error can't find the container with id 343e5005cdeb0ff6766f027efa57c372a04cd267a5631fbd060c9778fa310cbf Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.867928 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.867988 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.868003 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.868022 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.868036 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.971291 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.971361 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.971379 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.971407 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:19 crc kubenswrapper[4935]: I1005 06:53:19.971457 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:19Z","lastTransitionTime":"2025-10-05T06:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.047886 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" event={"ID":"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c","Type":"ContainerStarted","Data":"343e5005cdeb0ff6766f027efa57c372a04cd267a5631fbd060c9778fa310cbf"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.051566 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/0.log" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.056129 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.056309 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.078069 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.078121 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.078139 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.078168 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.078187 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.081517 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.103235 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.124878 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.143820 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.162991 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.181279 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.181343 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.181362 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.181388 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.181407 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.199624 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:18Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:53:18.101739 6206 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:53:18.101789 6206 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:53:18.101857 6206 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:53:18.101903 6206 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:53:18.101944 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:53:18.101972 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:53:18.101997 6206 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:53:18.101864 6206 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:53:18.101999 6206 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:53:18.102021 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:53:18.102032 6206 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:53:18.102062 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:53:18.102108 6206 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:53:18.102175 6206 factory.go:656] Stopping watch factory\\\\nI1005 06:53:18.102209 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:53:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.227260 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.242251 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.259635 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.280445 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.284354 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.284423 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.284448 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.284477 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.284498 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.299347 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.320329 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.342955 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.359132 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.383795 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.388711 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.388754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.388767 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.388789 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.388805 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.491921 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.491975 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.491986 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.492006 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.492016 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.573295 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-85tzm"] Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.573791 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:20 crc kubenswrapper[4935]: E1005 06:53:20.573858 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.588118 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.595265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.595497 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.595574 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.595665 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.595807 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.612473 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.634836 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.648784 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.698963 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.699020 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.699031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.699052 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.699063 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.699717 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.725259 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:18Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:53:18.101739 6206 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:53:18.101789 6206 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:53:18.101857 6206 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:53:18.101903 6206 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:53:18.101944 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:53:18.101972 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:53:18.101997 6206 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:53:18.101864 6206 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:53:18.101999 6206 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:53:18.102021 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:53:18.102032 6206 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:53:18.102062 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:53:18.102108 6206 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:53:18.102175 6206 factory.go:656] Stopping watch factory\\\\nI1005 06:53:18.102209 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:53:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.733919 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lrzj\" (UniqueName: \"kubernetes.io/projected/87a0c84e-d575-4468-af55-c814e305d34d-kube-api-access-9lrzj\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.734169 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.741839 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.754958 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.770139 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.776385 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.776496 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:20 crc kubenswrapper[4935]: E1005 06:53:20.776514 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:20 crc kubenswrapper[4935]: E1005 06:53:20.776653 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.776385 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:20 crc kubenswrapper[4935]: E1005 06:53:20.776735 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.787394 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.801810 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.801850 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.801860 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.801878 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.801906 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.802673 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.816242 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.835392 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.835474 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lrzj\" (UniqueName: \"kubernetes.io/projected/87a0c84e-d575-4468-af55-c814e305d34d-kube-api-access-9lrzj\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:20 crc kubenswrapper[4935]: E1005 06:53:20.835618 4935 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:20 crc kubenswrapper[4935]: E1005 06:53:20.835715 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs podName:87a0c84e-d575-4468-af55-c814e305d34d nodeName:}" failed. No retries permitted until 2025-10-05 06:53:21.335690439 +0000 UTC m=+35.218316899 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs") pod "network-metrics-daemon-85tzm" (UID: "87a0c84e-d575-4468-af55-c814e305d34d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.838111 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.855312 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.859946 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lrzj\" (UniqueName: \"kubernetes.io/projected/87a0c84e-d575-4468-af55-c814e305d34d-kube-api-access-9lrzj\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.866064 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.878837 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.904974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.905028 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.905038 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.905067 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:20 crc kubenswrapper[4935]: I1005 06:53:20.905081 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:20Z","lastTransitionTime":"2025-10-05T06:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.008271 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.008322 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.008332 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.008351 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.008363 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.065587 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/1.log" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.066258 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/0.log" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.069574 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6" exitCode=1 Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.069642 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.069725 4935 scope.go:117] "RemoveContainer" containerID="c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.070515 4935 scope.go:117] "RemoveContainer" containerID="c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6" Oct 05 06:53:21 crc kubenswrapper[4935]: E1005 06:53:21.070759 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.073293 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" event={"ID":"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c","Type":"ContainerStarted","Data":"66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.073367 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" event={"ID":"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c","Type":"ContainerStarted","Data":"f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.089119 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.109556 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:18Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:53:18.101739 6206 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:53:18.101789 6206 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:53:18.101857 6206 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:53:18.101903 6206 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:53:18.101944 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:53:18.101972 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:53:18.101997 6206 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:53:18.101864 6206 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:53:18.101999 6206 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:53:18.102021 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:53:18.102032 6206 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:53:18.102062 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:53:18.102108 6206 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:53:18.102175 6206 factory.go:656] Stopping watch factory\\\\nI1005 06:53:18.102209 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:53:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"message\\\":\\\"pi-operator-machine-webhook] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007842eeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.111182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.111230 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.111244 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.111266 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.111282 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.126555 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.139502 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.156585 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.173877 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.188858 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.206151 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.215100 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.215164 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.215187 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.215221 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.215243 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.224131 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.240725 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.260853 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.279006 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.293517 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.306373 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.318116 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.318172 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.318193 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.318222 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.318240 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.321571 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.334400 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.341883 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:21 crc kubenswrapper[4935]: E1005 06:53:21.342082 4935 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:21 crc kubenswrapper[4935]: E1005 06:53:21.342174 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs podName:87a0c84e-d575-4468-af55-c814e305d34d nodeName:}" failed. No retries permitted until 2025-10-05 06:53:22.342148337 +0000 UTC m=+36.224774827 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs") pod "network-metrics-daemon-85tzm" (UID: "87a0c84e-d575-4468-af55-c814e305d34d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.350349 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.368248 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.396887 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:18Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:53:18.101739 6206 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:53:18.101789 6206 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:53:18.101857 6206 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:53:18.101903 6206 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:53:18.101944 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:53:18.101972 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:53:18.101997 6206 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:53:18.101864 6206 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:53:18.101999 6206 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:53:18.102021 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:53:18.102032 6206 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:53:18.102062 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:53:18.102108 6206 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:53:18.102175 6206 factory.go:656] Stopping watch factory\\\\nI1005 06:53:18.102209 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:53:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"message\\\":\\\"pi-operator-machine-webhook] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007842eeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.422108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.422176 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.422195 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.422222 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.422244 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.423395 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.438263 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.460815 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.475351 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.488362 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.501754 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.517776 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.525625 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.525707 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.525733 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.525769 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.525815 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.541570 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.566702 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.582768 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.597527 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.615082 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.629232 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.629300 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.629320 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.629350 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.629369 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.632017 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.733100 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.733196 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.733210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.733265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.733283 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.837864 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.837974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.837996 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.838025 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.838046 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.941717 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.941809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.941840 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.941875 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:21 crc kubenswrapper[4935]: I1005 06:53:21.941924 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:21Z","lastTransitionTime":"2025-10-05T06:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.044254 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.044301 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.044312 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.044333 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.044347 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.077749 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/1.log" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.147466 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.147512 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.147524 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.147546 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.147559 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.250078 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.250148 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.250162 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.250185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.250200 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.352805 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.352855 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.352863 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.352879 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.352914 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.353413 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.353552 4935 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.353612 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs podName:87a0c84e-d575-4468-af55-c814e305d34d nodeName:}" failed. No retries permitted until 2025-10-05 06:53:24.353594973 +0000 UTC m=+38.236221433 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs") pod "network-metrics-daemon-85tzm" (UID: "87a0c84e-d575-4468-af55-c814e305d34d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.455005 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.455129 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:53:38.455101192 +0000 UTC m=+52.337727652 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.455242 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.455306 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.455412 4935 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.455463 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:38.455452021 +0000 UTC m=+52.338078481 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.455548 4935 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.455703 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:38.455670027 +0000 UTC m=+52.338296487 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.456591 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.456663 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.456689 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.456763 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.456792 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.556818 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.556884 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.557130 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.557154 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.557173 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.557246 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.557265 4935 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.557188 4935 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.557343 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:38.557320261 +0000 UTC m=+52.439946901 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.557501 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:53:38.557463705 +0000 UTC m=+52.440090175 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.558796 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.558852 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.558871 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.558928 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.558963 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.662317 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.662350 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.662368 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.662382 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.662392 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.765089 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.765155 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.765173 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.765199 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.765227 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.776565 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.776622 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.776644 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.776786 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.776840 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.777084 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.777288 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:22 crc kubenswrapper[4935]: E1005 06:53:22.777376 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.869548 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.869607 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.869625 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.869659 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.869679 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.973590 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.973686 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.973706 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.973737 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:22 crc kubenswrapper[4935]: I1005 06:53:22.973758 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:22Z","lastTransitionTime":"2025-10-05T06:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.077315 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.077415 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.077440 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.077476 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.077500 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.181276 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.181354 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.181372 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.181398 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.181418 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.285011 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.285108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.285133 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.285198 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.285218 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.387696 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.387773 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.387792 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.387820 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.387839 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.490857 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.490961 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.490979 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.491005 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.491024 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.594750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.594806 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.594823 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.594850 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.594869 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.699373 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.699487 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.699570 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.699606 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.699642 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.803606 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.803699 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.803724 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.803762 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.803786 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.907055 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.907128 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.907150 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.907178 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:23 crc kubenswrapper[4935]: I1005 06:53:23.907196 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:23Z","lastTransitionTime":"2025-10-05T06:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.010766 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.010816 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.010831 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.010851 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.010866 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.113632 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.113674 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.113684 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.113701 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.113715 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.218014 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.218113 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.218127 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.218148 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.218165 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.322112 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.322185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.322204 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.322235 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.322254 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.377687 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:24 crc kubenswrapper[4935]: E1005 06:53:24.377936 4935 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:24 crc kubenswrapper[4935]: E1005 06:53:24.378039 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs podName:87a0c84e-d575-4468-af55-c814e305d34d nodeName:}" failed. No retries permitted until 2025-10-05 06:53:28.378010706 +0000 UTC m=+42.260637206 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs") pod "network-metrics-daemon-85tzm" (UID: "87a0c84e-d575-4468-af55-c814e305d34d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.426006 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.426088 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.426113 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.426145 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.426174 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.528939 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.528986 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.529009 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.529034 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.529050 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.632318 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.632369 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.632386 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.632410 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.632425 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.735940 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.735992 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.736009 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.736032 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.736048 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.776629 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.776688 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:24 crc kubenswrapper[4935]: E1005 06:53:24.776837 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.776874 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.776958 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:24 crc kubenswrapper[4935]: E1005 06:53:24.777116 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:24 crc kubenswrapper[4935]: E1005 06:53:24.777287 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:24 crc kubenswrapper[4935]: E1005 06:53:24.777409 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.839331 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.839383 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.839405 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.839435 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.839459 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.942554 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.942624 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.942648 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.942679 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:24 crc kubenswrapper[4935]: I1005 06:53:24.942702 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:24Z","lastTransitionTime":"2025-10-05T06:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.045566 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.045629 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.045651 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.045676 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.045696 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.148753 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.148820 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.148847 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.148879 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.148934 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.252164 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.252247 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.252270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.252302 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.252328 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.355291 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.355357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.355374 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.355403 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.355423 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.459273 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.459411 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.459439 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.459470 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.459493 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.563525 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.563596 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.563613 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.563640 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.563660 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.667324 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.667405 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.667428 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.667460 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.667486 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.770454 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.770563 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.770582 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.770607 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.770625 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.874368 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.874432 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.874452 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.874483 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.874508 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.978190 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.978260 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.978278 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.978308 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:25 crc kubenswrapper[4935]: I1005 06:53:25.978329 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:25Z","lastTransitionTime":"2025-10-05T06:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.082764 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.082843 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.082880 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.082949 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.082976 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.186515 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.186585 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.186605 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.186631 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.186648 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.290712 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.290958 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.290986 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.291021 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.291044 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.394344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.394426 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.394461 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.394496 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.394535 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.497983 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.498047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.498065 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.498090 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.498111 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.601498 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.601562 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.601585 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.601617 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.601640 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.704727 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.704791 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.704801 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.704818 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.704832 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.776738 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.776747 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:26 crc kubenswrapper[4935]: E1005 06:53:26.776877 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.777023 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:26 crc kubenswrapper[4935]: E1005 06:53:26.777263 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.777342 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:26 crc kubenswrapper[4935]: E1005 06:53:26.777389 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:26 crc kubenswrapper[4935]: E1005 06:53:26.777511 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.794228 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.808360 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.808421 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.808439 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.808465 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.808483 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.808922 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.821437 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.834406 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.848803 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.876121 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.894359 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.911549 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.916845 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.916967 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.916990 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.917016 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.917035 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.920824 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.920867 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.920882 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.920923 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.920942 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.929629 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: E1005 06:53:26.935502 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.939507 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.939569 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.939586 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.939615 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.939636 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.954715 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c872b984e31e4626d2e47576496f6290fbebda57b0bb4186350902b91e3d6ab7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:18Z\\\",\\\"message\\\":\\\"sip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:53:18.101739 6206 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:53:18.101789 6206 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:53:18.101857 6206 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:53:18.101903 6206 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:53:18.101944 6206 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:53:18.101972 6206 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:53:18.101997 6206 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:53:18.101864 6206 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:53:18.101999 6206 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:53:18.102021 6206 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:53:18.102032 6206 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:53:18.102062 6206 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:53:18.102108 6206 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:53:18.102175 6206 factory.go:656] Stopping watch factory\\\\nI1005 06:53:18.102209 6206 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:53:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"message\\\":\\\"pi-operator-machine-webhook] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007842eeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: E1005 06:53:26.958512 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.965179 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.965254 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.965275 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.965310 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.965335 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:26 crc kubenswrapper[4935]: E1005 06:53:26.981221 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.984194 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.987216 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.987279 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.987310 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.987667 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:26 crc kubenswrapper[4935]: I1005 06:53:26.987689 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:26Z","lastTransitionTime":"2025-10-05T06:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:26.999868 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:27 crc kubenswrapper[4935]: E1005 06:53:27.004044 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.007779 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.007808 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.007819 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.007836 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.007849 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.013691 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:27 crc kubenswrapper[4935]: E1005 06:53:27.019822 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:27 crc kubenswrapper[4935]: E1005 06:53:27.019952 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.021407 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.021440 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.021454 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.021472 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.021484 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.029173 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.041520 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.052140 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.123099 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.123154 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.123167 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.123218 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.123240 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.225747 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.225815 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.225833 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.225858 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.225875 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.328540 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.328603 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.328621 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.328647 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.328665 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.431310 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.431429 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.431455 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.431491 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.431515 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.535196 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.535258 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.535276 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.535301 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.535318 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.639345 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.639494 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.639517 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.639545 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.639566 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.743687 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.743758 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.743781 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.743817 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.743842 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.847454 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.847519 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.847537 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.847594 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.847615 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.951595 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.951709 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.951736 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.951769 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:27 crc kubenswrapper[4935]: I1005 06:53:27.951789 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:27Z","lastTransitionTime":"2025-10-05T06:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.055459 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.055531 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.055556 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.055589 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.055618 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.158578 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.158632 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.158648 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.158672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.158689 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.262373 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.262444 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.262464 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.262488 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.262507 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.366007 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.366080 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.366102 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.366133 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.366156 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.423055 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:28 crc kubenswrapper[4935]: E1005 06:53:28.423270 4935 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:28 crc kubenswrapper[4935]: E1005 06:53:28.423382 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs podName:87a0c84e-d575-4468-af55-c814e305d34d nodeName:}" failed. No retries permitted until 2025-10-05 06:53:36.423354629 +0000 UTC m=+50.305981119 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs") pod "network-metrics-daemon-85tzm" (UID: "87a0c84e-d575-4468-af55-c814e305d34d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.470295 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.470365 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.470386 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.470412 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.470431 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.573736 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.573797 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.573816 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.573843 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.573864 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.677789 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.677855 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.677868 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.677928 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.677945 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.776486 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.776566 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.776573 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.776490 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:28 crc kubenswrapper[4935]: E1005 06:53:28.776762 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:28 crc kubenswrapper[4935]: E1005 06:53:28.776942 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:28 crc kubenswrapper[4935]: E1005 06:53:28.777145 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:28 crc kubenswrapper[4935]: E1005 06:53:28.777247 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.781231 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.781281 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.781302 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.781336 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.781356 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.884405 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.884462 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.884481 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.884508 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.884527 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.987331 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.987422 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.987446 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.987480 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:28 crc kubenswrapper[4935]: I1005 06:53:28.987501 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:28Z","lastTransitionTime":"2025-10-05T06:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.091285 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.091365 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.091383 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.091404 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.091420 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.194633 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.194695 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.194711 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.194736 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.194755 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.298948 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.299008 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.299025 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.299045 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.299056 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.402035 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.402114 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.402134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.402164 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.402185 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.505796 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.505869 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.505920 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.505950 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.505976 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.614982 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.615071 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.615090 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.615118 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.615138 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.718813 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.718948 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.718976 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.719008 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.719030 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.823135 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.823218 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.823242 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.823273 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.823296 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.926774 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.926852 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.926871 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.926930 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:29 crc kubenswrapper[4935]: I1005 06:53:29.926963 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:29Z","lastTransitionTime":"2025-10-05T06:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.030748 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.030812 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.030836 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.030884 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.030951 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.133787 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.133841 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.133854 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.133874 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.133886 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.237038 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.237134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.237157 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.237187 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.237207 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.340751 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.340819 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.340842 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.340873 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.340978 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.443953 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.444027 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.444050 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.444081 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.444103 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.547266 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.547328 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.547346 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.547370 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.547390 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.650684 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.650788 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.650817 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.650855 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.650876 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.755285 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.755359 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.755381 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.755415 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.755434 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.776762 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.776804 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.776858 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:30 crc kubenswrapper[4935]: E1005 06:53:30.777048 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.777135 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:30 crc kubenswrapper[4935]: E1005 06:53:30.777312 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:30 crc kubenswrapper[4935]: E1005 06:53:30.777480 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:30 crc kubenswrapper[4935]: E1005 06:53:30.777592 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.858515 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.858578 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.858592 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.858799 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.858812 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.962131 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.962210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.962227 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.962256 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:30 crc kubenswrapper[4935]: I1005 06:53:30.962275 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:30Z","lastTransitionTime":"2025-10-05T06:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.065927 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.066021 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.066047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.066086 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.066118 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.169060 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.169143 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.169166 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.169256 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.169299 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.273247 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.273337 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.273356 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.273423 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.273452 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.376948 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.377498 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.377519 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.377548 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.377568 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.481158 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.481217 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.481235 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.481261 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.481279 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.584826 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.584947 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.584974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.585002 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.585021 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.688420 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.688492 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.688510 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.688537 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.688557 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.791746 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.791813 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.791824 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.791845 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.791859 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.895188 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.895242 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.895259 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.895297 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.895323 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.999300 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.999411 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:31 crc kubenswrapper[4935]: I1005 06:53:31.999439 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:31.999475 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:31.999500 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:31Z","lastTransitionTime":"2025-10-05T06:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.103310 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.103396 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.103422 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.103456 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.103478 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.207886 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.207984 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.208008 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.208038 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.208056 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.311558 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.311625 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.311640 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.311659 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.311672 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.414735 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.414777 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.414786 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.414804 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.414815 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.517041 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.517087 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.517100 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.517119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.517168 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.619690 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.619724 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.619733 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.619750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.619762 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.722476 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.722557 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.722578 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.722607 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.722626 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.777011 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.777057 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.777068 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:32 crc kubenswrapper[4935]: E1005 06:53:32.777197 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.777292 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:32 crc kubenswrapper[4935]: E1005 06:53:32.777369 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:32 crc kubenswrapper[4935]: E1005 06:53:32.777447 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:32 crc kubenswrapper[4935]: E1005 06:53:32.777782 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.778092 4935 scope.go:117] "RemoveContainer" containerID="c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.795208 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.815244 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.825268 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.825318 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.825330 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.825347 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.825360 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.829835 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.848761 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.864251 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.882424 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.902219 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.922413 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.928210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.928246 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.928259 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.928277 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.928290 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:32Z","lastTransitionTime":"2025-10-05T06:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.935559 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.953960 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.983538 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"message\\\":\\\"pi-operator-machine-webhook] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007842eeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:32 crc kubenswrapper[4935]: I1005 06:53:32.998792 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.030523 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.030972 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.031021 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.031037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.031060 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.031077 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.048014 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.062490 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.076390 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.127413 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/1.log" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.131058 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.131276 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.133225 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.133272 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.133285 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.133304 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.133318 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.149326 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.165252 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.182255 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.201151 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.213768 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.235553 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.235857 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.235979 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.236051 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.236133 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.236597 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.247677 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.258790 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.277814 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"message\\\":\\\"pi-operator-machine-webhook] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007842eeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.291068 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.301313 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.311379 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.326483 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.338696 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.339076 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.339145 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.339157 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.339174 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.339189 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.350011 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.362682 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.441998 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.442573 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.442660 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.442765 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.442842 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.545603 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.545648 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.545658 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.545675 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.545684 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.648521 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.648577 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.648598 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.648623 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.648640 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.752507 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.752587 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.752607 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.752636 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.752656 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.856372 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.856434 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.856484 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.856511 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.856533 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.959736 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.959810 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.959835 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.959868 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:33 crc kubenswrapper[4935]: I1005 06:53:33.959938 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:33Z","lastTransitionTime":"2025-10-05T06:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.063325 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.063401 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.063424 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.063457 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.063479 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.144217 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/2.log" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.145669 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/1.log" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.151711 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4" exitCode=1 Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.151791 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.151859 4935 scope.go:117] "RemoveContainer" containerID="c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.154495 4935 scope.go:117] "RemoveContainer" containerID="e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4" Oct 05 06:53:34 crc kubenswrapper[4935]: E1005 06:53:34.154840 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.168387 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.168454 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.168471 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.168501 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.168521 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.176157 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.189242 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.208639 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.223694 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.263683 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.276047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.276115 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.276127 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.276146 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.276159 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.287819 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.303003 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.325235 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"message\\\":\\\"pi-operator-machine-webhook] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007842eeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.340515 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.350270 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.362081 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.376024 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.377776 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.377808 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.377818 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.377832 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.377841 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.389318 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.400679 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.412089 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.426712 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.480393 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.480440 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.480450 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.480464 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.480473 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.583043 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.583089 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.583107 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.583127 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.583139 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.686582 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.686633 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.686642 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.686659 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.686671 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.776507 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.776551 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.776570 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.776700 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:34 crc kubenswrapper[4935]: E1005 06:53:34.776974 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:34 crc kubenswrapper[4935]: E1005 06:53:34.777095 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:34 crc kubenswrapper[4935]: E1005 06:53:34.777256 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:34 crc kubenswrapper[4935]: E1005 06:53:34.777320 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.789487 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.789533 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.789542 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.789559 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.789570 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.893687 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.893754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.893773 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.893799 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.893818 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.997395 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.997482 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.997505 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.997533 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:34 crc kubenswrapper[4935]: I1005 06:53:34.997560 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:34Z","lastTransitionTime":"2025-10-05T06:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.101427 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.101466 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.101474 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.101488 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.101501 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.163020 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/2.log" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.207461 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.207522 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.207536 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.207560 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.207574 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.310743 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.311082 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.311198 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.311308 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.311408 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.414680 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.414756 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.414772 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.414794 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.414807 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.517964 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.518035 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.518049 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.518070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.518084 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.620656 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.620715 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.620735 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.620766 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.620790 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.724027 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.724086 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.724098 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.724120 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.724135 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.828919 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.828980 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.828994 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.829014 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.829026 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.932539 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.932598 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.932612 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.932633 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:35 crc kubenswrapper[4935]: I1005 06:53:35.932647 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:35Z","lastTransitionTime":"2025-10-05T06:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.035810 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.035873 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.035909 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.035937 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.035953 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.139431 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.139487 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.139499 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.139523 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.139538 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.243037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.243110 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.243129 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.243154 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.243172 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.346281 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.346344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.346361 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.346388 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.346406 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.449013 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.449081 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.449102 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.449128 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.449148 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.511962 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:36 crc kubenswrapper[4935]: E1005 06:53:36.512218 4935 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:36 crc kubenswrapper[4935]: E1005 06:53:36.512387 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs podName:87a0c84e-d575-4468-af55-c814e305d34d nodeName:}" failed. No retries permitted until 2025-10-05 06:53:52.512346196 +0000 UTC m=+66.394972696 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs") pod "network-metrics-daemon-85tzm" (UID: "87a0c84e-d575-4468-af55-c814e305d34d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.552071 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.552128 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.552145 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.552170 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.552195 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.656161 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.656201 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.656209 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.656240 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.656252 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.759439 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.759540 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.759571 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.759605 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.759631 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.777069 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.777124 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.777175 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.777193 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:36 crc kubenswrapper[4935]: E1005 06:53:36.777376 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:36 crc kubenswrapper[4935]: E1005 06:53:36.777475 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:36 crc kubenswrapper[4935]: E1005 06:53:36.777600 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:36 crc kubenswrapper[4935]: E1005 06:53:36.777704 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.803409 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.823310 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.844690 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.864189 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.864259 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.864278 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.864307 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.864326 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.866143 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.883768 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.905527 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.923323 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.951016 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92c5e259a92cbb255ed01c7d8f80473261156e8010e1d0f8d44c18a6ff4e4a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"message\\\":\\\"pi-operator-machine-webhook] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:machine-api-operator-machine-webhook-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007842eeb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{1 0 machine-webhook},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{api: clusterapi,k8s-app: controller,},ClusterIP:10.217.4.250,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.250],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.965094 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.966097 4935 scope.go:117] "RemoveContainer" containerID="e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4" Oct 05 06:53:36 crc kubenswrapper[4935]: E1005 06:53:36.966418 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.966930 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.966967 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.966976 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.966988 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.966997 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:36Z","lastTransitionTime":"2025-10-05T06:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.971051 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.982616 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:36 crc kubenswrapper[4935]: I1005 06:53:36.996664 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.009497 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.023696 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.038332 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.048885 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.065866 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.069803 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.069850 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.069910 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.069934 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.069946 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.079069 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.087357 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.099239 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.111058 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.120757 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.133343 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.143473 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.155389 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.172861 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.172914 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.172927 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.172943 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.172953 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.174006 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.188215 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.199179 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.216361 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.228952 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.228985 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.228994 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.229009 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.229019 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.229309 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.241735 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: E1005 06:53:37.242166 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.245145 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.245178 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.245187 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.245200 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.245212 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.254260 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: E1005 06:53:37.259800 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.262801 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.262826 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.262836 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.262855 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.262864 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.270770 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: E1005 06:53:37.281453 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.284644 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.284712 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.284735 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.284766 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.284792 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: E1005 06:53:37.303545 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.307048 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.307117 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.307140 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.307171 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.307196 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: E1005 06:53:37.319728 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:37 crc kubenswrapper[4935]: E1005 06:53:37.320090 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.322645 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.322705 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.322727 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.322748 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.322766 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.425305 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.425403 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.425423 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.425452 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.425470 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.528951 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.529374 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.529655 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.529928 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.530124 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.632830 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.632864 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.632873 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.632901 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.632911 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.735968 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.736079 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.736104 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.736137 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.736158 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.839252 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.839343 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.839364 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.839390 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.839411 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.942822 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.942857 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.942866 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.942878 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:37 crc kubenswrapper[4935]: I1005 06:53:37.942905 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:37Z","lastTransitionTime":"2025-10-05T06:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.045696 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.045763 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.045781 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.045807 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.045827 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.148750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.148802 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.148820 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.148845 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.148862 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.252110 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.252175 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.252192 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.252216 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.252236 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.355239 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.355311 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.355335 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.355392 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.355415 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.458157 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.458244 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.458265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.458293 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.458312 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.533992 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.534312 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.534383 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.534488 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:54:10.534440955 +0000 UTC m=+84.417067445 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.534505 4935 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.534622 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:54:10.534597179 +0000 UTC m=+84.417223669 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.534491 4935 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.534709 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:54:10.534694571 +0000 UTC m=+84.417321071 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.561146 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.561199 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.561217 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.561240 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.561258 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.635386 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.635442 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.635627 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.635649 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.635662 4935 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.635717 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.635758 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.635778 4935 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.635732 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:54:10.635710728 +0000 UTC m=+84.518337188 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.635880 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:54:10.635856572 +0000 UTC m=+84.518483072 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.664584 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.664647 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.664664 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.664689 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.664707 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.768678 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.768729 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.768740 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.768756 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.768768 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.777192 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.777239 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.777332 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.777361 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.777500 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.777614 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.777762 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:38 crc kubenswrapper[4935]: E1005 06:53:38.777872 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.871134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.871189 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.871200 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.871221 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.871235 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.974984 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.975048 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.975066 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.975091 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:38 crc kubenswrapper[4935]: I1005 06:53:38.975112 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:38Z","lastTransitionTime":"2025-10-05T06:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.078146 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.078212 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.078235 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.078267 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.078289 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.180330 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.180385 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.180402 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.180430 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.180454 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.282811 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.282929 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.282956 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.282996 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.283021 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.386431 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.386495 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.386514 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.386539 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.386557 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.489507 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.489613 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.489630 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.489658 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.489675 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.593004 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.593064 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.593082 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.593105 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.593122 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.696884 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.697025 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.697047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.697075 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.697097 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.801166 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.801231 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.801250 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.801277 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.801298 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.904706 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.904801 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.904821 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.904846 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:39 crc kubenswrapper[4935]: I1005 06:53:39.904864 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:39Z","lastTransitionTime":"2025-10-05T06:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.008498 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.008598 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.008616 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.008672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.008690 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.111853 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.111988 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.112009 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.112034 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.112082 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.215231 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.215307 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.215320 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.215344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.215357 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.319109 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.319214 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.319233 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.319259 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.319276 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.423047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.423130 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.423154 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.423189 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.423212 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.525492 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.525854 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.526081 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.526265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.526420 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.630305 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.630339 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.630347 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.630362 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.630372 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.733722 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.733762 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.733772 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.733789 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.733799 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.776964 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.777013 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.776974 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.777172 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:40 crc kubenswrapper[4935]: E1005 06:53:40.777158 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:40 crc kubenswrapper[4935]: E1005 06:53:40.777274 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:40 crc kubenswrapper[4935]: E1005 06:53:40.777312 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:40 crc kubenswrapper[4935]: E1005 06:53:40.777337 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.823372 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.836378 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.836419 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.836430 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.836446 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.836459 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.837017 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.837548 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.858163 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.878221 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.898078 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.908927 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.921986 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.935791 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.938940 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.938992 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.939009 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.939029 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.939041 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:40Z","lastTransitionTime":"2025-10-05T06:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.950945 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.964683 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:40 crc kubenswrapper[4935]: I1005 06:53:40.982299 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.002184 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:40Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.021127 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:41Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.034448 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:41Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.041270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.041344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.041366 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.041396 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.041416 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.052388 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:41Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.069007 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:41Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.084259 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:41Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.144028 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.144117 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.144137 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.144170 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.144192 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.246362 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.246420 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.246432 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.246451 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.246463 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.349847 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.349933 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.349945 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.349966 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.349978 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.452819 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.452880 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.452913 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.452934 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.452951 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.555236 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.555279 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.555292 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.555308 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.555320 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.656918 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.656991 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.657004 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.657022 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.657036 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.759020 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.759068 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.759078 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.759095 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.759106 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.861079 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.861153 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.861176 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.861205 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.861227 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.964057 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.964135 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.964153 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.964182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:41 crc kubenswrapper[4935]: I1005 06:53:41.964204 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:41Z","lastTransitionTime":"2025-10-05T06:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.067680 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.067746 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.067763 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.067788 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.067811 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.171309 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.171419 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.171444 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.171474 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.171494 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.274437 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.274504 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.274522 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.274547 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.274566 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.377571 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.377635 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.377655 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.377681 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.377698 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.480195 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.480245 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.480261 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.480281 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.480296 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.582968 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.583024 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.583043 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.583069 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.583088 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.686066 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.686110 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.686119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.686134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.686143 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.777118 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.777229 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.777289 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.777484 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:42 crc kubenswrapper[4935]: E1005 06:53:42.777462 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:42 crc kubenswrapper[4935]: E1005 06:53:42.777565 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:42 crc kubenswrapper[4935]: E1005 06:53:42.777668 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:42 crc kubenswrapper[4935]: E1005 06:53:42.777729 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.788263 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.788307 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.788319 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.788336 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.788349 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.890491 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.890549 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.890568 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.890591 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.890606 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.993179 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.993229 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.993243 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.993264 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:42 crc kubenswrapper[4935]: I1005 06:53:42.993278 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:42Z","lastTransitionTime":"2025-10-05T06:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.096027 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.096066 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.096078 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.096094 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.096105 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.198822 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.199111 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.199203 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.199320 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.199468 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.302495 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.302760 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.302868 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.303091 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.303195 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.406151 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.406197 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.406212 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.406236 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.406251 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.509399 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.509456 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.509473 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.509493 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.509507 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.613059 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.613135 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.613162 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.613198 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.613221 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.716406 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.716474 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.716494 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.716518 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.716536 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.819951 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.819996 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.820010 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.820031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.820044 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.923665 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.923739 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.923882 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.924001 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:43 crc kubenswrapper[4935]: I1005 06:53:43.924044 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:43Z","lastTransitionTime":"2025-10-05T06:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.027476 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.027549 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.027569 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.027594 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.027612 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.130270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.130328 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.130341 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.130360 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.130375 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.233096 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.233138 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.233150 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.233167 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.233179 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.335962 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.335990 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.335998 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.336013 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.336022 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.439283 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.439326 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.439334 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.439348 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.439357 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.542473 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.542539 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.542556 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.542579 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.542598 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.645296 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.645361 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.645382 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.645404 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.645418 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.749301 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.749355 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.749371 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.749393 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.749406 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.777036 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.777079 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.777140 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:44 crc kubenswrapper[4935]: E1005 06:53:44.777202 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.777215 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:44 crc kubenswrapper[4935]: E1005 06:53:44.777310 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:44 crc kubenswrapper[4935]: E1005 06:53:44.777467 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:44 crc kubenswrapper[4935]: E1005 06:53:44.777564 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.851879 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.851949 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.851960 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.851978 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.851991 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.956065 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.956160 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.956187 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.956224 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:44 crc kubenswrapper[4935]: I1005 06:53:44.956249 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:44Z","lastTransitionTime":"2025-10-05T06:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.059672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.059762 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.059776 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.059794 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.059809 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.163411 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.163496 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.163516 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.163542 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.163564 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.266669 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.266734 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.266768 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.266800 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.266822 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.370391 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.370464 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.370486 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.370519 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.370543 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.474007 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.474086 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.474106 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.474133 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.474153 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.577295 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.577357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.577373 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.577400 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.577417 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.680224 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.680269 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.680293 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.680316 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.680330 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.783353 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.783417 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.783458 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.783485 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.783504 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.887232 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.887303 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.887323 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.887353 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.887373 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.990732 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.990781 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.990799 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.990824 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:45 crc kubenswrapper[4935]: I1005 06:53:45.990842 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:45Z","lastTransitionTime":"2025-10-05T06:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.094428 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.094481 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.094492 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.094513 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.094526 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.197556 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.197619 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.197631 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.197649 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.197662 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.300536 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.300607 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.300624 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.300651 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.300669 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.403951 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.404010 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.404027 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.404053 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.404071 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.506625 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.506705 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.506742 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.506775 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.506797 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.610686 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.610760 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.610787 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.610811 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.610829 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.715618 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.715655 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.715691 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.715717 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.715730 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.776218 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.776238 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.776370 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.776437 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:46 crc kubenswrapper[4935]: E1005 06:53:46.776468 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:46 crc kubenswrapper[4935]: E1005 06:53:46.776604 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:46 crc kubenswrapper[4935]: E1005 06:53:46.776722 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:46 crc kubenswrapper[4935]: E1005 06:53:46.776804 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.796007 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.813301 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.819697 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.819738 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.819754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.819774 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.819791 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.829310 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.844760 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.858548 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.879114 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.893591 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.905325 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.917550 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.922000 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.922046 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.922057 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.922075 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.922085 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:46Z","lastTransitionTime":"2025-10-05T06:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.930616 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.943451 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.959473 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.972875 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.983329 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:46 crc kubenswrapper[4935]: I1005 06:53:46.995564 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:46Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.006567 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.018098 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.024622 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.024655 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.024667 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.024685 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.024697 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.127612 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.127652 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.127664 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.127678 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.127689 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.229425 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.229461 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.229471 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.229484 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.229494 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.332108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.332164 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.332177 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.332195 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.332208 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.434711 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.434746 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.434755 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.434771 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.434781 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.473122 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.473182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.473195 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.473211 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.473220 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: E1005 06:53:47.492958 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.497552 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.497598 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.497611 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.497628 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.497639 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: E1005 06:53:47.508607 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.511935 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.511957 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.511966 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.511976 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.511984 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: E1005 06:53:47.525256 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.528662 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.528690 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.528698 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.528708 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.528718 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: E1005 06:53:47.543462 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.547080 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.547291 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.547625 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.547791 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.547963 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: E1005 06:53:47.563843 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:47 crc kubenswrapper[4935]: E1005 06:53:47.564142 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.565588 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.565613 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.565621 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.565635 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.565644 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.669499 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.669549 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.669566 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.669588 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.669605 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.772668 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.772715 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.772730 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.772748 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.772759 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.876686 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.876785 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.876840 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.876866 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.876886 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.979032 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.979117 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.979142 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.979173 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:47 crc kubenswrapper[4935]: I1005 06:53:47.979194 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:47Z","lastTransitionTime":"2025-10-05T06:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.081965 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.082011 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.082021 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.082038 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.082048 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.184903 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.184961 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.184972 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.184995 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.185009 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.287809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.287872 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.287929 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.287954 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.287966 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.390809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.390861 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.390870 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.390913 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.390930 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.493799 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.493840 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.493851 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.493867 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.493878 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.596543 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.596622 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.596644 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.596679 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.596704 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.704104 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.704212 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.704233 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.704265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.704286 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.776292 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.776411 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.776442 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:48 crc kubenswrapper[4935]: E1005 06:53:48.776503 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.776713 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:48 crc kubenswrapper[4935]: E1005 06:53:48.776711 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:48 crc kubenswrapper[4935]: E1005 06:53:48.776797 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:48 crc kubenswrapper[4935]: E1005 06:53:48.777038 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.806962 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.807007 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.807021 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.807042 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.807059 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.910024 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.910072 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.910085 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.910102 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:48 crc kubenswrapper[4935]: I1005 06:53:48.910115 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:48Z","lastTransitionTime":"2025-10-05T06:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.013705 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.013757 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.013766 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.013784 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.013794 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.116623 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.116699 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.116715 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.116742 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.116764 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.219765 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.219836 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.219853 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.219877 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.219930 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.322628 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.322721 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.322777 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.322809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.322828 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.426127 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.426175 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.426185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.426203 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.426214 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.529526 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.529597 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.529626 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.529657 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.529681 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.632448 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.632521 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.632538 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.632565 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.632584 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.738632 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.738711 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.738730 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.738756 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.738777 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.778330 4935 scope.go:117] "RemoveContainer" containerID="e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4" Oct 05 06:53:49 crc kubenswrapper[4935]: E1005 06:53:49.778707 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.841950 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.842035 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.842060 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.842091 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.842115 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.946137 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.946226 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.946246 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.946277 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:49 crc kubenswrapper[4935]: I1005 06:53:49.946299 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:49Z","lastTransitionTime":"2025-10-05T06:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.049696 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.049823 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.049846 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.049876 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.049930 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.153651 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.153743 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.153761 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.153790 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.153809 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.257489 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.257575 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.257601 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.257650 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.257675 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.360471 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.360563 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.360589 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.360626 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.360655 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.464354 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.464406 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.464415 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.464436 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.464448 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.567648 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.567733 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.567754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.567790 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.567814 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.670230 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.670306 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.670326 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.670354 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.670372 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.773246 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.773338 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.773363 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.773400 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.773426 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.776591 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.776681 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.776733 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.776814 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:50 crc kubenswrapper[4935]: E1005 06:53:50.776806 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:50 crc kubenswrapper[4935]: E1005 06:53:50.777031 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:50 crc kubenswrapper[4935]: E1005 06:53:50.777239 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:50 crc kubenswrapper[4935]: E1005 06:53:50.777340 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.876628 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.876687 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.876698 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.876718 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.876730 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.979555 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.979618 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.979633 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.979653 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:50 crc kubenswrapper[4935]: I1005 06:53:50.979667 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:50Z","lastTransitionTime":"2025-10-05T06:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.083422 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.083475 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.083486 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.083501 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.083512 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.186370 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.186432 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.186448 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.186466 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.186479 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.289783 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.289846 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.289866 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.289924 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.289951 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.393069 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.393108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.393120 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.393139 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.393152 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.496702 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.496766 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.496778 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.496799 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.496812 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.599405 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.599480 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.599504 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.599538 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.599563 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.702626 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.702685 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.702706 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.702735 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.702755 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.805578 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.805617 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.805627 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.805643 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.805653 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.909061 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.909151 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.909175 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.909211 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:51 crc kubenswrapper[4935]: I1005 06:53:51.909246 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:51Z","lastTransitionTime":"2025-10-05T06:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.012490 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.012540 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.012558 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.012585 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.012606 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.115129 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.115191 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.115209 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.115238 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.115260 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.219425 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.219520 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.219543 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.219573 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.219600 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.322749 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.322805 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.322823 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.322846 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.322861 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.426106 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.426175 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.426198 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.426225 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.426245 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.529319 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.529390 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.529402 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.529428 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.529442 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.590909 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:52 crc kubenswrapper[4935]: E1005 06:53:52.591198 4935 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:52 crc kubenswrapper[4935]: E1005 06:53:52.591333 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs podName:87a0c84e-d575-4468-af55-c814e305d34d nodeName:}" failed. No retries permitted until 2025-10-05 06:54:24.5912963 +0000 UTC m=+98.473922930 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs") pod "network-metrics-daemon-85tzm" (UID: "87a0c84e-d575-4468-af55-c814e305d34d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.632812 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.632878 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.632923 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.632952 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.632968 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.735566 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.735606 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.735671 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.735690 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.735704 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.776419 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.776522 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.776422 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:52 crc kubenswrapper[4935]: E1005 06:53:52.776690 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.776442 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:52 crc kubenswrapper[4935]: E1005 06:53:52.776564 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:52 crc kubenswrapper[4935]: E1005 06:53:52.776824 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:52 crc kubenswrapper[4935]: E1005 06:53:52.776879 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.838824 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.838885 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.838946 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.838979 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.839000 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.941670 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.941760 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.941778 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.941809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:52 crc kubenswrapper[4935]: I1005 06:53:52.941831 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:52Z","lastTransitionTime":"2025-10-05T06:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.045393 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.045484 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.045497 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.045514 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.045527 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.148620 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.148750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.148776 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.148810 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.148834 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.251704 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.251821 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.251841 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.251869 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.251917 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.354948 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.355013 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.355034 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.355063 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.355086 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.458078 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.458146 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.458165 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.458191 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.458210 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.561449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.561510 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.561529 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.561551 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.561569 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.664142 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.664210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.664231 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.664260 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.664280 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.766973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.767029 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.767040 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.767057 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.767069 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.869966 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.870031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.870048 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.870073 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.870092 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.974204 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.974272 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.974293 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.974322 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:53 crc kubenswrapper[4935]: I1005 06:53:53.974341 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:53Z","lastTransitionTime":"2025-10-05T06:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.077629 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.077707 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.077726 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.077763 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.077783 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.181044 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.181106 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.181123 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.181147 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.181168 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.284102 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.284166 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.284186 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.284211 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.284228 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.386672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.386739 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.386756 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.386780 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.386799 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.489330 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.489395 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.489406 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.489425 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.489436 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.592837 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.592916 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.592930 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.592956 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.592968 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.695872 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.696145 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.696227 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.696311 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.696384 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.777176 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.777230 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.777177 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.777177 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:54 crc kubenswrapper[4935]: E1005 06:53:54.777326 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:54 crc kubenswrapper[4935]: E1005 06:53:54.777374 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:54 crc kubenswrapper[4935]: E1005 06:53:54.777461 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:54 crc kubenswrapper[4935]: E1005 06:53:54.777524 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.798776 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.798838 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.798849 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.798864 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.798874 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.902178 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.902231 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.902243 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.902261 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:54 crc kubenswrapper[4935]: I1005 06:53:54.902273 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:54Z","lastTransitionTime":"2025-10-05T06:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.005419 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.005487 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.005511 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.005546 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.005569 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.108258 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.108292 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.108301 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.108316 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.108325 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.211511 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.211587 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.211605 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.211626 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.211642 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.314992 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.315035 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.315045 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.315060 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.315075 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.417371 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.417410 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.417422 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.417440 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.417454 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.520215 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.520258 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.520276 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.520297 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.520312 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.623392 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.623440 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.623459 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.623481 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.623498 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.725998 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.726233 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.726324 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.726408 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.726475 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.830326 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.830733 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.830948 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.831146 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.831332 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.933969 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.934037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.934059 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.934089 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:55 crc kubenswrapper[4935]: I1005 06:53:55.934109 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:55Z","lastTransitionTime":"2025-10-05T06:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.036915 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.036968 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.036982 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.037000 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.037012 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.139490 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.139553 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.139566 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.139585 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.139986 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.242868 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.242943 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.242958 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.242981 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.242996 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.345668 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.345703 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.345711 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.345724 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.345733 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.448837 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.448954 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.448975 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.449002 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.449020 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.552517 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.552579 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.552596 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.552618 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.552632 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.655099 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.655144 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.655156 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.655174 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.655186 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813565 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813605 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813615 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813608 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813645 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813571 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:56 crc kubenswrapper[4935]: E1005 06:53:56.813709 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813574 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813661 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: E1005 06:53:56.813782 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:56 crc kubenswrapper[4935]: E1005 06:53:56.813845 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.813832 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: E1005 06:53:56.813923 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.826692 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.840006 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.849722 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.860534 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.869899 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.879922 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.894455 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.906971 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.915960 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.915997 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.916007 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.916023 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.916032 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:56Z","lastTransitionTime":"2025-10-05T06:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.920937 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.931901 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.948952 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.967792 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:56 crc kubenswrapper[4935]: I1005 06:53:56.983956 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.000970 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:56Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.015222 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.019517 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.019570 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.019581 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.019598 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.019608 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.037500 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.049476 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.121258 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.121336 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.121355 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.121381 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.121403 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.223651 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.223694 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.223705 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.223720 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.223733 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.238701 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/0.log" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.238760 4935 generic.go:334] "Generic (PLEG): container finished" podID="2b9ff491-4f10-4f1b-822b-a8c46871ce60" containerID="7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea" exitCode=1 Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.238797 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-687xm" event={"ID":"2b9ff491-4f10-4f1b-822b-a8c46871ce60","Type":"ContainerDied","Data":"7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.239326 4935 scope.go:117] "RemoveContainer" containerID="7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.252932 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.265860 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.278307 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.288876 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.300462 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.309221 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.320663 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.327474 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.327518 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.327530 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.327547 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.327559 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.332351 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:56Z\\\",\\\"message\\\":\\\"2025-10-05T06:53:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87\\\\n2025-10-05T06:53:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87 to /host/opt/cni/bin/\\\\n2025-10-05T06:53:11Z [verbose] multus-daemon started\\\\n2025-10-05T06:53:11Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:53:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.353353 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.368036 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.379118 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.393148 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.408925 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.419007 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.428256 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.429843 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.429874 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.429898 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.429913 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.429922 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.440418 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.452692 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.533341 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.533396 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.533410 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.533432 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.533446 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.636348 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.636403 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.636421 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.636445 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.636464 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.739347 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.739395 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.739406 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.739425 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.739436 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.766031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.766102 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.766125 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.766156 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.766179 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: E1005 06:53:57.779073 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.783490 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.783524 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.783536 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.783552 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.783565 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: E1005 06:53:57.794203 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.801305 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.801356 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.801380 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.801397 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.801475 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: E1005 06:53:57.817100 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.820377 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.820436 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.820451 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.820470 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.820483 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: E1005 06:53:57.833733 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.836967 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.836991 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.837001 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.837014 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.837024 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: E1005 06:53:57.849060 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:57Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:57 crc kubenswrapper[4935]: E1005 06:53:57.849172 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.850520 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.850540 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.850548 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.850559 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.850568 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.953029 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.953091 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.953109 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.953134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:57 crc kubenswrapper[4935]: I1005 06:53:57.953153 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:57Z","lastTransitionTime":"2025-10-05T06:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.055672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.055765 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.055788 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.055817 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.055840 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.158345 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.158375 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.158385 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.158402 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.158414 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.246079 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/0.log" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.246174 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-687xm" event={"ID":"2b9ff491-4f10-4f1b-822b-a8c46871ce60","Type":"ContainerStarted","Data":"7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.260778 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.260813 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.260822 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.260839 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.260851 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.276788 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.294317 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.304962 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.318610 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.334168 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.352162 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.363221 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.363259 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.363270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.363290 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.363304 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.365423 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.381903 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:56Z\\\",\\\"message\\\":\\\"2025-10-05T06:53:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87\\\\n2025-10-05T06:53:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87 to /host/opt/cni/bin/\\\\n2025-10-05T06:53:11Z [verbose] multus-daemon started\\\\n2025-10-05T06:53:11Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:53:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.394009 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.410858 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.424986 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.437081 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.449448 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.458541 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.465759 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.465806 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.465818 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.465834 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.465845 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.469157 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.479447 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.488770 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:53:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.568343 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.568394 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.568403 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.568418 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.568430 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.671492 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.671534 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.671549 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.671567 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.671582 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.774036 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.774102 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.774115 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.774136 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.774151 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.776295 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:53:58 crc kubenswrapper[4935]: E1005 06:53:58.776424 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.776440 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.776546 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:53:58 crc kubenswrapper[4935]: E1005 06:53:58.776576 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.776440 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:53:58 crc kubenswrapper[4935]: E1005 06:53:58.776758 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:53:58 crc kubenswrapper[4935]: E1005 06:53:58.776785 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.876202 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.876252 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.876263 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.876280 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.876295 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.979696 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.979806 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.979820 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.979837 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:58 crc kubenswrapper[4935]: I1005 06:53:58.979847 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:58Z","lastTransitionTime":"2025-10-05T06:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.082578 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.082663 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.082681 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.082736 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.082763 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.186065 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.186102 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.186112 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.186130 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.186139 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.289142 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.289185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.289193 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.289237 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.289246 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.392210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.392267 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.392282 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.392303 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.392678 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.496422 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.496477 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.496490 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.496508 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.496519 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.600938 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.600997 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.601015 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.601038 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.601055 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.702846 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.702911 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.702928 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.702946 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.702958 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.806211 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.806268 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.806279 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.806294 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.806304 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.908834 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.908868 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.908878 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.908940 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:53:59 crc kubenswrapper[4935]: I1005 06:53:59.908954 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:53:59Z","lastTransitionTime":"2025-10-05T06:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.011186 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.011228 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.011242 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.011262 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.011277 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.113380 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.113419 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.113443 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.113459 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.113469 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.216175 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.216222 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.216236 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.216254 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.216267 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.319985 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.320042 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.320055 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.320076 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.320089 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.422980 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.423018 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.423029 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.423046 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.423056 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.525309 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.525339 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.525349 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.525365 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.525374 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.627923 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.627974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.627988 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.628006 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.628021 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.730732 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.730774 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.730786 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.730803 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.730813 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.776662 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:00 crc kubenswrapper[4935]: E1005 06:54:00.776780 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.776962 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:00 crc kubenswrapper[4935]: E1005 06:54:00.777031 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.777149 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:00 crc kubenswrapper[4935]: E1005 06:54:00.777209 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.777856 4935 scope.go:117] "RemoveContainer" containerID="e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.776626 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:00 crc kubenswrapper[4935]: E1005 06:54:00.780052 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.834005 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.834057 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.834068 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.834083 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.834094 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.936403 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.936446 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.936466 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.936490 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:00 crc kubenswrapper[4935]: I1005 06:54:00.936507 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:00Z","lastTransitionTime":"2025-10-05T06:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.039402 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.039440 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.039449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.039465 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.039475 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.141943 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.141990 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.142001 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.142023 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.142035 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.245578 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.245625 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.245642 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.245662 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.245675 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.257494 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/2.log" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.264173 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.264727 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.295300 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:54:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.317939 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.346763 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.348311 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.348352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.348363 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.348380 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.348392 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.374920 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.391431 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.404159 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.413553 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.423850 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:56Z\\\",\\\"message\\\":\\\"2025-10-05T06:53:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87\\\\n2025-10-05T06:53:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87 to /host/opt/cni/bin/\\\\n2025-10-05T06:53:11Z [verbose] multus-daemon started\\\\n2025-10-05T06:53:11Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:53:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.435770 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.448195 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.450626 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.450654 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.450662 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.450676 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.450688 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.464490 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.475878 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.486716 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.498515 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.508883 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.518393 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.526926 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:01Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.553708 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.553761 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.553776 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.553798 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.553810 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.656541 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.656598 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.656620 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.656644 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.656663 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.759161 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.759208 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.759221 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.759241 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.759254 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.861644 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.861710 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.861731 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.861757 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.861781 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.963858 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.963921 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.963935 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.963960 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:01 crc kubenswrapper[4935]: I1005 06:54:01.963985 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:01Z","lastTransitionTime":"2025-10-05T06:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.067010 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.067050 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.067064 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.067086 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.067100 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.169459 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.169543 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.169568 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.169618 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.169642 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.271618 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/3.log" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.271832 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.271880 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.271933 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.271970 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.271994 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.272487 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/2.log" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.276032 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" exitCode=1 Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.276078 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.276142 4935 scope.go:117] "RemoveContainer" containerID="e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.277300 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 06:54:02 crc kubenswrapper[4935]: E1005 06:54:02.277650 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.293608 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.310280 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.328554 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.344983 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.361011 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.375130 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.375171 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.375185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.375205 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.375220 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.380728 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.398584 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.416683 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.434063 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.449659 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:56Z\\\",\\\"message\\\":\\\"2025-10-05T06:53:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87\\\\n2025-10-05T06:53:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87 to /host/opt/cni/bin/\\\\n2025-10-05T06:53:11Z [verbose] multus-daemon started\\\\n2025-10-05T06:53:11Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:53:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.470133 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc5eeea6e512e19a791db3433121b0596529105269c214d44c3fed899540b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:33Z\\\",\\\"message\\\":\\\"IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:53:33.690305 6555 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1005 06:53:33.690414 6555 services_controller.go:444] Built service openshift-dns-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690429 6555 services_controller.go:445] Built service openshift-dns-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI1005 06:53:33.690432 6555 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 892.524µs\\\\nI1005 06:53:33.690456 6555 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1005 06:53:33.690457 6555 services_controller.go:451] Built service openshift-dns-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns-operator/metrics\\\\\\\"}, Opts:services.LBOp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:54:01Z\\\",\\\"message\\\":\\\"6:54:01.731068 6910 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:54:01.732876 6910 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:54:01.732940 6910 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:54:01.732961 6910 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:54:01.733021 6910 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:54:01.733022 6910 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:54:01.733039 6910 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:54:01.733031 6910 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:54:01.733046 6910 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:54:01.733052 6910 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:54:01.733071 6910 factory.go:656] Stopping watch factory\\\\nI1005 06:54:01.733083 6910 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:54:01.733098 6910 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:54:01.733107 6910 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:54:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.477593 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.477654 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.477669 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.477731 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.477745 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.484288 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.493510 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.505623 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.517694 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.528263 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.539072 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:02Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.580185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.580243 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.580260 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.580284 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.580301 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.683108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.683165 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.683179 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.683198 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.683213 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.776539 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:02 crc kubenswrapper[4935]: E1005 06:54:02.776677 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.776743 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:02 crc kubenswrapper[4935]: E1005 06:54:02.776786 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.777018 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:02 crc kubenswrapper[4935]: E1005 06:54:02.777089 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.777304 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:02 crc kubenswrapper[4935]: E1005 06:54:02.777455 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.785475 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.785525 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.785542 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.785568 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.785586 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.889771 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.889915 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.889931 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.889956 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.889969 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.994317 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.994357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.994372 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.994388 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:02 crc kubenswrapper[4935]: I1005 06:54:02.994400 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:02Z","lastTransitionTime":"2025-10-05T06:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.097015 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.097069 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.097083 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.097103 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.097116 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.201285 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.201339 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.201352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.201373 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.201387 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.283028 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/3.log" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.289374 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 06:54:03 crc kubenswrapper[4935]: E1005 06:54:03.289718 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.305761 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.305852 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.305869 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.305928 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.305949 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.312914 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:56Z\\\",\\\"message\\\":\\\"2025-10-05T06:53:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87\\\\n2025-10-05T06:53:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87 to /host/opt/cni/bin/\\\\n2025-10-05T06:53:11Z [verbose] multus-daemon started\\\\n2025-10-05T06:53:11Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:53:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.380526 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:54:01Z\\\",\\\"message\\\":\\\"6:54:01.731068 6910 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:54:01.732876 6910 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:54:01.732940 6910 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:54:01.732961 6910 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:54:01.733021 6910 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:54:01.733022 6910 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:54:01.733039 6910 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:54:01.733031 6910 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:54:01.733046 6910 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:54:01.733052 6910 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:54:01.733071 6910 factory.go:656] Stopping watch factory\\\\nI1005 06:54:01.733083 6910 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:54:01.733098 6910 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:54:01.733107 6910 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:54:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.409011 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.409061 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.409072 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.409089 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.409099 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.409921 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.430733 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.452629 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.469570 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.485972 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.504208 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.513090 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.513166 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.513184 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.513210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.513256 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.522226 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.542708 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.570170 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.590546 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.610217 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.622962 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.623063 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.623102 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.624348 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.624413 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.631851 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.652562 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.672969 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.690940 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:03Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.727756 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.727797 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.727807 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.727831 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.727845 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.832661 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.832754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.832780 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.832817 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.832842 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.936969 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.937068 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.937139 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.937182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:03 crc kubenswrapper[4935]: I1005 06:54:03.937211 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:03Z","lastTransitionTime":"2025-10-05T06:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.040680 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.040740 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.040757 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.040782 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.040800 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.144833 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.144964 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.144989 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.145022 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.145051 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.249553 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.249682 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.249714 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.249758 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.249784 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.354220 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.354289 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.354316 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.354344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.354362 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.457768 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.457860 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.457881 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.457952 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.457976 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.560796 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.560966 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.560981 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.560996 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.561004 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.664613 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.664692 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.664712 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.664742 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.664762 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.768299 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.768394 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.768409 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.768426 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.768437 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.776833 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.776845 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.776933 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:04 crc kubenswrapper[4935]: E1005 06:54:04.776949 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.777037 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:04 crc kubenswrapper[4935]: E1005 06:54:04.777090 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:04 crc kubenswrapper[4935]: E1005 06:54:04.777200 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:04 crc kubenswrapper[4935]: E1005 06:54:04.777726 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.792697 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.872701 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.872754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.872768 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.872788 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.872802 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.977068 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.977139 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.977163 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.977192 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:04 crc kubenswrapper[4935]: I1005 06:54:04.977212 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:04Z","lastTransitionTime":"2025-10-05T06:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.080763 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.080822 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.080840 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.080865 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.080884 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.184339 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.184416 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.184428 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.184448 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.184460 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.288129 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.288185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.288203 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.288231 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.288250 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.391974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.392052 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.392079 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.392114 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.392136 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.495690 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.495805 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.495835 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.495939 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.496004 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.599277 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.599414 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.599435 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.599499 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.599519 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.703607 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.703678 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.703696 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.703723 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.703744 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.808498 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.808585 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.808603 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.809115 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.809199 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.913391 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.913467 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.913507 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.913542 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:05 crc kubenswrapper[4935]: I1005 06:54:05.913568 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:05Z","lastTransitionTime":"2025-10-05T06:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.017518 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.017585 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.017603 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.017627 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.017645 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.120862 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.120954 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.120971 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.120995 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.121013 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.223957 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.224023 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.224043 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.224066 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.224080 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.327619 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.328274 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.328297 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.328331 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.328356 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.432031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.432071 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.432080 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.432099 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.432112 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.535655 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.535712 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.535734 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.535762 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.535782 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.639057 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.639130 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.639148 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.639176 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.639198 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.742152 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.742221 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.742240 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.742270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.742289 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.777229 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.777230 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.777326 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:06 crc kubenswrapper[4935]: E1005 06:54:06.777612 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.777878 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:06 crc kubenswrapper[4935]: E1005 06:54:06.778204 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:06 crc kubenswrapper[4935]: E1005 06:54:06.778338 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:06 crc kubenswrapper[4935]: E1005 06:54:06.778513 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.801449 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.825107 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.843789 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.845556 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.845608 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.845627 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.845652 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.845672 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.861810 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.887744 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.910502 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.931366 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.948691 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.949081 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.949313 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.949473 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.949628 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:06Z","lastTransitionTime":"2025-10-05T06:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.951543 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.969985 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:06 crc kubenswrapper[4935]: I1005 06:54:06.996015 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:56Z\\\",\\\"message\\\":\\\"2025-10-05T06:53:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87\\\\n2025-10-05T06:53:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87 to /host/opt/cni/bin/\\\\n2025-10-05T06:53:11Z [verbose] multus-daemon started\\\\n2025-10-05T06:53:11Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:53:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:06Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.030880 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:54:01Z\\\",\\\"message\\\":\\\"6:54:01.731068 6910 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:54:01.732876 6910 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:54:01.732940 6910 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:54:01.732961 6910 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:54:01.733021 6910 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:54:01.733022 6910 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:54:01.733039 6910 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:54:01.733031 6910 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:54:01.733046 6910 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:54:01.733052 6910 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:54:01.733071 6910 factory.go:656] Stopping watch factory\\\\nI1005 06:54:01.733083 6910 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:54:01.733098 6910 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:54:01.733107 6910 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:54:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.052719 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.052803 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.052824 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.052856 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.052885 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.059729 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.079165 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.102387 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.122279 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.141055 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97017e9f-b44d-4c78-822d-dc6158c841b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af3c608e5fb2b5e5244d6cf09e93f739f8c8a6171452e5a621e860de7ad186a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c71942182bd653c831680f6c9d9ef0b2e58a5747904364e2ce3a6a5d9786e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c71942182bd653c831680f6c9d9ef0b2e58a5747904364e2ce3a6a5d9786e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.156619 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.156698 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.156716 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.156754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.156777 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.164154 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.193806 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:07Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.259764 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.260165 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.260383 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.260582 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.260784 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.364637 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.364676 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.364687 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.364704 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.364716 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.468574 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.468621 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.468633 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.468655 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.468669 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.571052 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.571116 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.571135 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.571161 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.571178 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.675291 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.675371 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.675398 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.675427 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.675470 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.778449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.778531 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.778550 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.778574 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.778591 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.881601 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.881671 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.881689 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.881715 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.881735 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.985363 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.985415 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.985426 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.985449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:07 crc kubenswrapper[4935]: I1005 06:54:07.985461 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:07Z","lastTransitionTime":"2025-10-05T06:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.089776 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.089848 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.089866 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.089920 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.089939 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.093266 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.093314 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.093327 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.093345 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.093358 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.114105 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.120851 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.120923 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.120938 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.120958 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.120972 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.141527 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.146981 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.147047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.147099 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.147127 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.147146 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.168124 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.174148 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.174224 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.174243 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.174271 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.174291 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.196624 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.203182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.203253 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.203272 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.203301 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.203323 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.224061 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.224312 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.227419 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.227475 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.227496 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.227530 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.227551 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.331300 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.331407 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.331431 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.331464 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.331486 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.436937 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.437010 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.437031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.437057 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.437076 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.540362 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.540781 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.540988 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.541144 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.541361 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.645535 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.645621 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.645645 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.645679 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.645701 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.749881 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.749996 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.750016 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.750054 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.750077 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.776802 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.776846 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.777203 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.777363 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.777396 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.777506 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.777731 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:08 crc kubenswrapper[4935]: E1005 06:54:08.777838 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.854571 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.854651 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.854672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.854702 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.854724 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.958778 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.958877 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.958927 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.958957 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:08 crc kubenswrapper[4935]: I1005 06:54:08.958981 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:08Z","lastTransitionTime":"2025-10-05T06:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.062483 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.062732 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.062882 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.063083 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.063234 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.166300 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.166356 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.166378 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.166409 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.166430 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.270019 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.270149 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.270174 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.270208 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.270231 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.373587 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.373633 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.373652 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.373680 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.373703 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.477305 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.477352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.477386 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.477404 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.477416 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.580749 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.580812 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.580831 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.580857 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.580875 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.684855 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.684935 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.684948 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.684968 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.684979 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.788290 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.788361 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.788387 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.788417 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.788446 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.891496 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.891543 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.891560 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.891584 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.891602 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.994500 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.994564 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.994583 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.994605 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:09 crc kubenswrapper[4935]: I1005 06:54:09.994622 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:09Z","lastTransitionTime":"2025-10-05T06:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.098224 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.098290 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.098313 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.098347 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.098369 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.201726 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.201804 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.201831 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.201861 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.201881 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.304774 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.304852 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.304875 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.304941 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.304965 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.408331 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.408478 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.408546 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.408575 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.408600 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.511609 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.511650 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.511662 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.511678 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.511691 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.606009 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.606148 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.606265 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.606279 4935 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.606333 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.606302363 +0000 UTC m=+148.488928853 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.606388 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.606360084 +0000 UTC m=+148.488986584 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.606418 4935 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.606509 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.606464857 +0000 UTC m=+148.489091357 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.615282 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.615335 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.615352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.615376 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.615396 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.707436 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.707991 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.707661 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.708434 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.708497 4935 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.708093 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.708677 4935 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.708702 4935 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.708595 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.708581324 +0000 UTC m=+148.591207784 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.708768 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.708749358 +0000 UTC m=+148.591375818 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.717835 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.717872 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.717881 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.717914 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.717924 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.776824 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.776876 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.776965 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.777546 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.777609 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.776972 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.777684 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:10 crc kubenswrapper[4935]: E1005 06:54:10.778376 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.821265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.821361 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.821387 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.821418 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.821440 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.925242 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.925310 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.925327 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.925352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:10 crc kubenswrapper[4935]: I1005 06:54:10.925372 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:10Z","lastTransitionTime":"2025-10-05T06:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.027991 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.028048 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.028072 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.028100 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.028123 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.131227 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.131305 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.131329 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.131358 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.131380 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.234738 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.234786 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.234802 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.234823 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.234841 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.337426 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.337511 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.337535 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.337565 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.337585 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.440828 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.440932 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.440955 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.440982 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.441000 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.544236 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.544305 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.544326 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.544352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.544369 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.647057 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.647123 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.647147 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.647179 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.647201 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.750679 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.750738 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.750760 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.750790 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.750810 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.854644 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.854740 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.854759 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.854784 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.854806 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.959041 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.959119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.959138 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.959172 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:11 crc kubenswrapper[4935]: I1005 06:54:11.959193 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:11Z","lastTransitionTime":"2025-10-05T06:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.062573 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.062635 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.062654 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.062678 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.062697 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.166605 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.166689 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.166710 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.166741 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.166762 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.270282 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.270382 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.270407 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.270446 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.270478 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.374460 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.374515 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.374536 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.374564 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.374582 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.478345 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.478412 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.478432 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.478462 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.478483 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.582299 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.582392 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.582416 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.582450 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.582469 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.685368 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.685455 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.685481 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.685518 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.685584 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.777081 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.777121 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.777106 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.777268 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:12 crc kubenswrapper[4935]: E1005 06:54:12.777482 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:12 crc kubenswrapper[4935]: E1005 06:54:12.777595 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:12 crc kubenswrapper[4935]: E1005 06:54:12.777733 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:12 crc kubenswrapper[4935]: E1005 06:54:12.777943 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.788622 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.788671 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.788695 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.788724 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.788748 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.892459 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.892943 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.893206 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.893433 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.893607 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.997039 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.997104 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.997122 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.997148 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:12 crc kubenswrapper[4935]: I1005 06:54:12.997167 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:12Z","lastTransitionTime":"2025-10-05T06:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.100850 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.100953 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.100978 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.101011 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.101039 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.203772 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.203822 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.203836 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.203856 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.203877 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.313027 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.313456 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.313640 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.313825 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.314103 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.418304 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.418379 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.418398 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.418426 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.418443 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.521406 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.521870 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.522118 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.522179 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.522210 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.625478 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.625556 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.625572 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.625595 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.625616 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.728883 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.728995 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.729018 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.729047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.729068 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.832073 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.832162 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.832200 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.832237 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.832259 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.935606 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.935677 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.935701 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.935732 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:13 crc kubenswrapper[4935]: I1005 06:54:13.935758 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:13Z","lastTransitionTime":"2025-10-05T06:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.038357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.038445 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.038470 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.038500 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.038517 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.141731 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.141814 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.141830 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.141878 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.141940 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.244288 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.244371 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.244396 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.244425 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.244443 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.347731 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.347866 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.347918 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.347947 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.347966 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.450670 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.450717 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.450735 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.450758 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.450777 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.553639 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.553750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.553774 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.553799 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.553818 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.656296 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.656347 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.656363 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.656389 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.656406 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.759757 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.759857 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.759877 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.759949 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.759968 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.777007 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.777104 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.777130 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:14 crc kubenswrapper[4935]: E1005 06:54:14.777196 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:14 crc kubenswrapper[4935]: E1005 06:54:14.777347 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.777376 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:14 crc kubenswrapper[4935]: E1005 06:54:14.777517 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:14 crc kubenswrapper[4935]: E1005 06:54:14.777673 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.863191 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.863284 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.863302 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.863329 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.863379 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.967264 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.967333 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.967352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.967377 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:14 crc kubenswrapper[4935]: I1005 06:54:14.967396 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:14Z","lastTransitionTime":"2025-10-05T06:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.071539 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.071603 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.071621 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.071646 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.071663 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.174288 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.174346 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.174365 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.174395 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.174414 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.277302 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.277357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.277371 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.277393 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.277408 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.380437 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.380511 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.380534 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.380565 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.380589 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.483379 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.483469 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.483487 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.483512 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.483529 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.586291 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.586325 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.586339 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.586357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.586370 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.688629 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.688728 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.688752 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.688807 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.688827 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.777960 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 06:54:15 crc kubenswrapper[4935]: E1005 06:54:15.778344 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.791740 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.791792 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.791808 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.791832 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.791847 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.895867 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.895974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.895998 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.896031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.896052 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.999423 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.999490 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.999508 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.999533 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:15 crc kubenswrapper[4935]: I1005 06:54:15.999549 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:15Z","lastTransitionTime":"2025-10-05T06:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.101806 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.101865 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.101876 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.101918 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.101930 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.205099 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.205185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.205209 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.205237 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.205258 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.308657 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.308706 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.308723 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.308745 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.308760 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.411664 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.412069 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.412223 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.412391 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.412522 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.515466 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.515545 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.515569 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.515601 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.515623 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.618821 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.618880 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.618931 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.618965 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.619016 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.722278 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.722365 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.722383 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.722407 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.722425 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.776211 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:16 crc kubenswrapper[4935]: E1005 06:54:16.776379 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.776406 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.776468 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.776550 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:16 crc kubenswrapper[4935]: E1005 06:54:16.776690 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:16 crc kubenswrapper[4935]: E1005 06:54:16.776797 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:16 crc kubenswrapper[4935]: E1005 06:54:16.776886 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.797640 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.816977 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.825582 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.825632 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.825649 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.825672 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.825689 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.839100 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:56Z\\\",\\\"message\\\":\\\"2025-10-05T06:53:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87\\\\n2025-10-05T06:53:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87 to /host/opt/cni/bin/\\\\n2025-10-05T06:53:11Z [verbose] multus-daemon started\\\\n2025-10-05T06:53:11Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:53:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.870649 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:54:01Z\\\",\\\"message\\\":\\\"6:54:01.731068 6910 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:54:01.732876 6910 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:54:01.732940 6910 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:54:01.732961 6910 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:54:01.733021 6910 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:54:01.733022 6910 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:54:01.733039 6910 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:54:01.733031 6910 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:54:01.733046 6910 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:54:01.733052 6910 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:54:01.733071 6910 factory.go:656] Stopping watch factory\\\\nI1005 06:54:01.733083 6910 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:54:01.733098 6910 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:54:01.733107 6910 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:54:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.900886 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.919947 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.928571 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.928609 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.928653 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.928674 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.928687 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:16Z","lastTransitionTime":"2025-10-05T06:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.940390 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.955277 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.969922 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:16 crc kubenswrapper[4935]: I1005 06:54:16.981173 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97017e9f-b44d-4c78-822d-dc6158c841b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af3c608e5fb2b5e5244d6cf09e93f739f8c8a6171452e5a621e860de7ad186a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c71942182bd653c831680f6c9d9ef0b2e58a5747904364e2ce3a6a5d9786e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c71942182bd653c831680f6c9d9ef0b2e58a5747904364e2ce3a6a5d9786e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.000799 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.023886 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.032828 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.032944 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.032965 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.033383 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.033427 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.042939 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.062863 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.079167 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.097826 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.115467 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.128146 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:17Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.136635 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.136723 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.136745 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.136772 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.136792 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.239702 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.239759 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.239777 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.239802 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.239820 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.342750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.342795 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.342805 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.342822 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.342833 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.445049 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.445079 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.445090 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.445105 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.445117 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.547143 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.547622 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.547641 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.547675 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.547693 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.650610 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.650662 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.650674 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.650693 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.650705 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.754087 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.754134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.754146 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.754165 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.754178 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.857149 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.857260 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.857278 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.857302 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.857318 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.960181 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.960248 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.960265 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.960288 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:17 crc kubenswrapper[4935]: I1005 06:54:17.960305 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:17Z","lastTransitionTime":"2025-10-05T06:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.063407 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.063787 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.064025 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.064248 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.064399 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.168070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.168289 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.168377 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.168443 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.168504 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.271657 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.271939 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.272067 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.272119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.272131 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.367834 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.368271 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.368365 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.368462 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.368550 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.380779 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.384499 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.384552 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.384569 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.384590 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.384608 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.403015 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.407179 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.407343 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.407432 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.407517 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.407588 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.418829 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.422587 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.422611 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.422624 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.422638 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.422648 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.434203 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.438006 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.438114 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.438238 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.438358 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.438541 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.456520 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.456670 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.458237 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.458284 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.458301 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.458326 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.458341 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.561814 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.561926 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.561974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.562000 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.562017 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.665044 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.665119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.665135 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.665157 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.665176 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.768182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.768232 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.768255 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.768283 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.768303 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.777185 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.777186 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.777250 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.777330 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.777480 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.777598 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.777776 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:18 crc kubenswrapper[4935]: E1005 06:54:18.778647 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.871416 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.871736 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.871958 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.872125 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.872274 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.975065 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.975125 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.975142 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.975168 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:18 crc kubenswrapper[4935]: I1005 06:54:18.975186 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:18Z","lastTransitionTime":"2025-10-05T06:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.078419 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.078978 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.079123 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.079243 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.079349 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.182536 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.182872 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.183060 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.183254 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.183384 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.286855 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.286985 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.287013 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.287052 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.287080 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.390098 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.390198 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.390214 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.390233 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.390245 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.494155 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.494217 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.494229 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.494249 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.494259 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.597749 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.597805 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.597817 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.597834 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.597846 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.701931 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.702134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.702160 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.702218 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.702239 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.806130 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.806186 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.806207 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.806235 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.806257 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.910548 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.910618 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.910638 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.910665 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:19 crc kubenswrapper[4935]: I1005 06:54:19.910685 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:19Z","lastTransitionTime":"2025-10-05T06:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.014659 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.014726 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.014745 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.014775 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.014794 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.118612 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.118710 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.118758 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.118786 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.118834 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.222472 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.222546 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.222568 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.222599 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.222620 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.325834 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.326230 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.326294 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.326352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.326380 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.429838 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.429973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.430014 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.430050 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.430075 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.533281 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.533340 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.533357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.533383 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.533402 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.636103 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.636181 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.636200 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.636230 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.636257 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.739415 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.739498 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.739521 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.739553 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.739573 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.777099 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.777211 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.777300 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:20 crc kubenswrapper[4935]: E1005 06:54:20.777298 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:20 crc kubenswrapper[4935]: E1005 06:54:20.777440 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.777618 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:20 crc kubenswrapper[4935]: E1005 06:54:20.777643 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:20 crc kubenswrapper[4935]: E1005 06:54:20.777696 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.843241 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.843299 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.843313 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.843339 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.843352 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.947651 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.947722 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.947740 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.947764 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:20 crc kubenswrapper[4935]: I1005 06:54:20.947780 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:20Z","lastTransitionTime":"2025-10-05T06:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.051070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.051127 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.051143 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.051167 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.051189 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.159047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.159121 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.159140 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.159167 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.159194 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.262935 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.263330 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.263552 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.263727 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.263957 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.367570 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.367632 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.367652 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.367676 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.367694 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.471430 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.471505 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.471523 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.471549 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.471566 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.574765 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.575213 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.575398 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.575603 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.575761 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.706591 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.707011 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.707222 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.707397 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.707549 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.816860 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.816945 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.816961 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.816988 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.817003 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.920381 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.920453 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.920469 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.920494 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:21 crc kubenswrapper[4935]: I1005 06:54:21.920516 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:21Z","lastTransitionTime":"2025-10-05T06:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.023232 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.023295 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.023311 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.023337 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.023354 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.126723 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.128051 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.128226 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.128400 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.128586 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.232087 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.232119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.232131 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.232147 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.232158 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.335258 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.335644 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.335827 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.336193 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.336355 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.439681 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.439730 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.439746 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.439769 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.439786 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.542935 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.543011 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.543031 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.543056 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.543075 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.646291 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.646729 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.646973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.647189 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.647345 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.750167 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.750232 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.750249 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.750278 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.750296 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.776525 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.776611 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.776679 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.776880 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:22 crc kubenswrapper[4935]: E1005 06:54:22.776858 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:22 crc kubenswrapper[4935]: E1005 06:54:22.777037 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:22 crc kubenswrapper[4935]: E1005 06:54:22.777145 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:22 crc kubenswrapper[4935]: E1005 06:54:22.777223 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.853904 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.853986 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.854003 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.854023 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.854035 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.957421 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.957667 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.957863 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.957996 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:22 crc kubenswrapper[4935]: I1005 06:54:22.958081 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:22Z","lastTransitionTime":"2025-10-05T06:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.061372 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.061410 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.061422 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.061439 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.061452 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.164656 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.164707 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.164723 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.164749 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.164766 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.267438 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.267519 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.267539 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.267564 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.267580 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.370289 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.370618 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.370710 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.370821 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.370931 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.474156 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.474212 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.474227 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.474255 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.474271 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.577078 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.577128 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.577141 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.577162 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.577178 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.680396 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.680449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.680461 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.680480 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.680492 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.783828 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.783956 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.783979 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.784004 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.784025 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.887615 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.887662 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.887678 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.887707 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.887721 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.991183 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.991235 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.991248 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.991268 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:23 crc kubenswrapper[4935]: I1005 06:54:23.991281 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:23Z","lastTransitionTime":"2025-10-05T06:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.094595 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.094642 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.094654 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.094673 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.094687 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.197967 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.198045 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.198069 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.198099 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.198121 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.301352 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.301449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.301486 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.301518 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.301616 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.403747 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.403809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.403826 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.403852 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.403875 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.506306 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.506337 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.506346 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.506358 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.506368 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.609332 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.609388 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.609406 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.609428 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.609441 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.669831 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:24 crc kubenswrapper[4935]: E1005 06:54:24.670118 4935 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:54:24 crc kubenswrapper[4935]: E1005 06:54:24.670210 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs podName:87a0c84e-d575-4468-af55-c814e305d34d nodeName:}" failed. No retries permitted until 2025-10-05 06:55:28.6701851 +0000 UTC m=+162.552811600 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs") pod "network-metrics-daemon-85tzm" (UID: "87a0c84e-d575-4468-af55-c814e305d34d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.711988 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.712037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.712052 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.712073 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.712086 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.776416 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.776460 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.776439 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.776590 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:24 crc kubenswrapper[4935]: E1005 06:54:24.776838 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:24 crc kubenswrapper[4935]: E1005 06:54:24.777571 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:24 crc kubenswrapper[4935]: E1005 06:54:24.778306 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:24 crc kubenswrapper[4935]: E1005 06:54:24.778978 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.801549 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.814794 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.814863 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.814882 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.814934 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.814953 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.918500 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.918545 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.918555 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.918572 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:24 crc kubenswrapper[4935]: I1005 06:54:24.918583 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:24Z","lastTransitionTime":"2025-10-05T06:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.021361 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.021456 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.021474 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.021497 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.021511 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.124859 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.124973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.124993 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.125016 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.125033 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.227853 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.227911 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.227922 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.227936 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.227944 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.333057 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.333100 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.333112 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.333129 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.333141 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.435791 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.435856 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.435874 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.435929 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.435948 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.538405 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.538450 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.538461 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.538477 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.538488 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.641397 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.641444 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.641456 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.641477 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.641490 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.744715 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.744754 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.744762 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.744780 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.744788 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.847660 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.847722 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.847744 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.847773 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.847793 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.950519 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.950600 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.950626 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.950658 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:25 crc kubenswrapper[4935]: I1005 06:54:25.950682 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:25Z","lastTransitionTime":"2025-10-05T06:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.053161 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.053210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.053223 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.053240 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.053254 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.155858 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.155974 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.155993 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.156016 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.156033 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.259649 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.259735 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.259760 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.259788 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.259809 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.363080 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.363141 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.363158 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.363182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.363201 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.466062 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.466122 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.466140 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.466162 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.466180 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.568799 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.568872 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.568932 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.568963 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.568986 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.671930 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.672008 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.672037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.672070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.672095 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.775372 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.775430 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.775450 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.775475 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.775494 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.776830 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.776987 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:26 crc kubenswrapper[4935]: E1005 06:54:26.777151 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.777439 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.777507 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:26 crc kubenswrapper[4935]: E1005 06:54:26.777601 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:26 crc kubenswrapper[4935]: E1005 06:54:26.777834 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:26 crc kubenswrapper[4935]: E1005 06:54:26.778206 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.779498 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 06:54:26 crc kubenswrapper[4935]: E1005 06:54:26.779772 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.802924 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e561041-9f5d-4fa0-9d9f-7f8e8001a010\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed93f0f101628de50c9a5063d2b08495f126fe076deb74fd6f7ac7b97fc4fe36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f23abd0f18bf95480abb33abcccc3e0117a5be9d9bc2293cf483ee7f3f97ee52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4b359858a2f67cb5dbe8a7146e70ebe2e092bc63bfeb128d3beb7a6392aa2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bb8189411d4b5999e6627434ea58ee203b7755b9797fcffd4ecf796717a587e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c36896e50fae5df884c5a13be14a2ca689233184e8ba6fc76f95826ee8bfddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9374586a384472fd723385827cdc9268e30d02def3740c44e7a4e56abbdd592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9374586a384472fd723385827cdc9268e30d02def3740c44e7a4e56abbdd592\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ae9cc68210f7d97c95a79f044f979b65a11be9828c9ce22456d497e5cd25975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ae9cc68210f7d97c95a79f044f979b65a11be9828c9ce22456d497e5cd25975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d45c1c61dd37f5b09210065acc7f2c3ebdbc9d8dc611088cca7dd98acb743e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45c1c61dd37f5b09210065acc7f2c3ebdbc9d8dc611088cca7dd98acb743e32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.823665 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.843718 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e64695de052ec256dc0869fe163fe8fccd008bfed34725ade1b95b29af629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.857389 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5h5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc6754de-5445-4fe9-8707-2bca9b3e955a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0cc8659481621b66b7e04ebabfdffda7ffd873df291e5959f3b25c16df3b8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcsgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5h5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.877779 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-687xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b9ff491-4f10-4f1b-822b-a8c46871ce60\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:53:56Z\\\",\\\"message\\\":\\\"2025-10-05T06:53:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87\\\\n2025-10-05T06:53:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dbcd3063-98b6-43fe-a2b1-55141c02de87 to /host/opt/cni/bin/\\\\n2025-10-05T06:53:11Z [verbose] multus-daemon started\\\\n2025-10-05T06:53:11Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:53:56Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-56m5n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-687xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.880772 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.881313 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.881501 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.881640 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.881752 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.911867 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"019e7dde-1935-4b4c-8312-99f52b6d22fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:54:01Z\\\",\\\"message\\\":\\\"6:54:01.731068 6910 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:54:01.732865 6910 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:54:01.732876 6910 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:54:01.732940 6910 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:54:01.732961 6910 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:54:01.733021 6910 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:54:01.733022 6910 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:54:01.733039 6910 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:54:01.733031 6910 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:54:01.733046 6910 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1005 06:54:01.733052 6910 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1005 06:54:01.733071 6910 factory.go:656] Stopping watch factory\\\\nI1005 06:54:01.733083 6910 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:54:01.733098 6910 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:54:01.733107 6910 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:54:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzplz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ns4c5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.933194 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f724e021-2e89-4192-bf08-76b1a2033a04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b077e131c1d611943b32ef51b83922a9e8e43a449f24de6603baaab4ad2a05c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5813dbd040c25b101daa163883a679ed29552c8dbbe08fbb018e0106624cdfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab67f224f101022d3f9cc938cebd26e5e04997a8c887398c73d6e522f61a2ec9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6bda61b76f88431fe6aa366eb0b6146e61e345facca1f1dcb36b4dc96e5ecf6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42cb78850e4ab4f7e3c6481cd6209e8c9e17606a2400004c8c07c975856538fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:53:00Z\\\",\\\"message\\\":\\\"W1005 06:52:49.987804 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 06:52:49.988117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759647169 cert, and key in /tmp/serving-cert-3340276075/serving-signer.crt, /tmp/serving-cert-3340276075/serving-signer.key\\\\nI1005 06:52:50.377307 1 observer_polling.go:159] Starting file observer\\\\nW1005 06:52:50.380454 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 06:52:50.380776 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:52:50.385662 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3340276075/tls.crt::/tmp/serving-cert-3340276075/tls.key\\\\\\\"\\\\nF1005 06:53:00.832602 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://554ffd6f5a70e1de4143f4df6c4a86fe6e8e614ff8cbf206e4ed90cc63b9abff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://217a76959b7eca91ccb28b46a28f56e6a707dd0afd54bbe73e91fac0b1fca80b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.950216 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gth9x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"551313e5-e4b6-4462-86c7-c6b7087fba96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858c1adfccbb54b6a99ebe7f0a0a4baddb2a3ad7cf9a2c723e857b0356cef63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wlzv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:06Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gth9x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.969847 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5233d4b50b50e8ba04ef36917cb523848acfe51181072608e32a048429c0c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.986575 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.986643 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.986670 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.986702 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.986728 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:26Z","lastTransitionTime":"2025-10-05T06:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:26 crc kubenswrapper[4935]: I1005 06:54:26.986879 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80b8cb86d8630b53bb91763d2b475c30a6ec0b100e8f78472658d4732f60798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ab948b345aebcc8c4dce251633ccf00e7b6f2591650a9988fdd1d022a1b536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.003931 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.016875 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62823424-a6c9-4733-b50a-6f3ecb4a55a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06333f38854a561524cd28cfb4f271c74a827bc2b77861ac2523d88523bfa25e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7s48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-j8cwh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.029074 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cb1942fa-f3fe-4aa9-99e3-09e12525ab6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4cce49be42f22647af3111e6ab2abbece2d148ea91381d1872b8c49119df11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66ddbb9e295f461f0c089da6ac5bfdb353981fc54d89d29ef91e2c4adb11cc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-znjdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.039580 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97017e9f-b44d-4c78-822d-dc6158c841b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af3c608e5fb2b5e5244d6cf09e93f739f8c8a6171452e5a621e860de7ad186a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c71942182bd653c831680f6c9d9ef0b2e58a5747904364e2ce3a6a5d9786e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c71942182bd653c831680f6c9d9ef0b2e58a5747904364e2ce3a6a5d9786e81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.053684 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71c27212-34a9-4b23-a554-fe58a07cacaf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69514dda44b10ef1cc65df1966c6806e0ab9f9604f40c57c22e8046f2c7b8a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92230b309815ad9bd8875b76304ca0b6b2a31e014266dd83afdf2fba697f1b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013cc4c8c2b7cee56845a46cc514ccc691bc6aae73d7ba7e28f103a50fa6dfbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce34acfa96e5331ce07ff70c07128eda08cc0455bce225001a913aa6a0ac6753\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.074670 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e242b9a-8af9-4ddc-9c51-27d1432c80fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0a7812e8e773831937c17cfd736215b7f77199bc80abff6fc09655e9ba40986\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7caa274c02b4bea7e7f57e8276a70204ea169046092e9ec50e85221d494c2b4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d4969d0a9639553fc1b6937ebca88b76372824d5c1ac6d4c6252b063de1d13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61216021f5652a2eb024f6726a868dac26abefd2216b3755be944376038379eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aed663bdcc8a1ee2287bf39232bf3d560b33dde75b1881367b3d13e1ad983771\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9045b521a707bf2365dd6eaf7b7bc0893ad86526fed95fdafc919e736351cc62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b175a19c9e772ed3569916129552bdfb7cbbcbc2f7cb411cfe3ac9a8032dbd6a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:53:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:53:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cz2rj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.091086 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.091308 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.091433 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.091563 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.091683 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.093147 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03411977-cabd-4b84-b737-3a3d891b7797\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6935ca9a1a5d5ab374c33a1f61e440476077ca20723b854a4cf79b7da99d81d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://299a8524c6c7765dc0a10afcb95be69a97020ffb793ab7cce351c8683a6f83c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbeec814e2f81ad075214a660ecb4ab460dc10019a94f388ace2f4d99de25003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:52:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7595ccb2b467efd7c85437c5b50232fe005e52b62105518a95bf6d6705ffb1e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:52:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:52:47Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:52:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.113489 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.128138 4935 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-85tzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a0c84e-d575-4468-af55-c814e305d34d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9lrzj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:53:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-85tzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.195453 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.195495 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.195507 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.195523 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.195535 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.297983 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.298034 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.298047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.298065 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.298078 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.400125 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.400209 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.400228 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.400249 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.400266 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.503338 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.503389 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.503406 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.503431 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.503452 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.606915 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.606961 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.606980 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.607002 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.607019 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.709452 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.710005 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.710448 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.710714 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.711381 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.814982 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.815049 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.815073 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.815106 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.815128 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.918300 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.918376 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.918396 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.918422 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:27 crc kubenswrapper[4935]: I1005 06:54:27.918439 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:27Z","lastTransitionTime":"2025-10-05T06:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.020540 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.020601 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.020623 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.020655 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.020682 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.123839 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.123905 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.123922 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.123942 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.123955 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.227614 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.227663 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.227674 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.227692 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.227706 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.331347 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.331407 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.331425 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.331449 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.331466 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.433879 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.433941 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.433954 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.433973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.433984 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.537163 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.537266 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.537290 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.537321 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.537343 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.639413 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.639542 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.639560 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.639577 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.639589 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.719809 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.719860 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.719873 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.719903 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.719916 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.736032 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.741152 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.741213 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.741237 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.741261 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.741285 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.761698 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.767287 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.767335 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.767362 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.767388 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.767410 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.776722 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.776926 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.776949 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.776997 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.777033 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.777119 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.777327 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.777466 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.794816 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.800713 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.800777 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.800796 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.800821 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.800842 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.822696 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.828208 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.828446 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.828572 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.828717 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.828801 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.848525 4935 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:54:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b377934a-1ade-44d6-a53c-ef84970bcbd9\\\",\\\"systemUUID\\\":\\\"dddde716-0390-449f-afc3-a07540401a49\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:54:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:54:28 crc kubenswrapper[4935]: E1005 06:54:28.848744 4935 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.850707 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.850751 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.850770 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.850794 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.850812 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.954724 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.955062 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.955159 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.955270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:28 crc kubenswrapper[4935]: I1005 06:54:28.955352 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:28Z","lastTransitionTime":"2025-10-05T06:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.058030 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.058105 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.058119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.058138 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.058159 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.163487 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.163567 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.163586 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.163618 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.163639 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.267186 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.267233 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.267242 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.267260 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.267275 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.370784 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.370855 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.370868 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.370904 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.370917 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.473517 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.473575 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.473593 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.473619 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.473637 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.576498 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.576558 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.576580 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.576605 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.576622 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.679070 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.679396 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.679465 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.679544 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.679615 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.782344 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.782411 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.782437 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.782469 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.782490 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.885823 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.886044 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.886082 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.886114 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.886139 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.989094 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.989164 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.989185 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.989214 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:29 crc kubenswrapper[4935]: I1005 06:54:29.989235 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:29Z","lastTransitionTime":"2025-10-05T06:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.091981 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.092019 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.092028 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.092046 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.092056 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.195479 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.195525 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.195536 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.195553 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.195565 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.298138 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.298203 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.298226 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.298366 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.298396 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.401263 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.401341 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.401368 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.401398 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.401419 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.504054 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.504097 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.504113 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.504129 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.504141 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.607348 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.607418 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.607440 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.607467 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.607488 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.710437 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.710464 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.710472 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.710484 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.710492 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.776468 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:30 crc kubenswrapper[4935]: E1005 06:54:30.776575 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.776654 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.776679 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:30 crc kubenswrapper[4935]: E1005 06:54:30.776817 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.776849 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:30 crc kubenswrapper[4935]: E1005 06:54:30.776981 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:30 crc kubenswrapper[4935]: E1005 06:54:30.777124 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.815281 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.815371 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.815398 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.815433 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.815472 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.919584 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.919637 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.919650 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.919669 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:30 crc kubenswrapper[4935]: I1005 06:54:30.919682 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:30Z","lastTransitionTime":"2025-10-05T06:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.023430 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.023498 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.023557 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.023588 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.023611 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.126644 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.126723 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.126742 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.126771 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.126789 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.230338 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.230399 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.230423 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.230454 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.230475 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.333973 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.334062 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.334097 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.334128 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.334149 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.437359 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.437554 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.437577 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.437602 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.437620 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.540806 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.540964 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.540982 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.541006 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.541024 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.643971 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.644028 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.644047 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.644072 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.644089 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.747022 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.747091 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.747101 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.747115 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.747126 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.850773 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.850823 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.850847 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.850875 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.850932 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.954458 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.954539 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.954559 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.954592 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:31 crc kubenswrapper[4935]: I1005 06:54:31.954618 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:31Z","lastTransitionTime":"2025-10-05T06:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.058091 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.058121 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.058134 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.058147 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.058156 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.160574 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.160628 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.160645 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.160675 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.160694 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.264155 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.264233 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.264254 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.264287 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.264314 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.367730 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.367790 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.367813 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.367840 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.367858 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.471210 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.471272 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.471293 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.471322 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.471341 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.574760 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.574825 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.574843 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.574868 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.574886 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.677635 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.677699 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.677720 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.677745 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.677763 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.777204 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.777242 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.777438 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.777475 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:32 crc kubenswrapper[4935]: E1005 06:54:32.777642 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:32 crc kubenswrapper[4935]: E1005 06:54:32.777725 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:32 crc kubenswrapper[4935]: E1005 06:54:32.778168 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:32 crc kubenswrapper[4935]: E1005 06:54:32.778233 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.780357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.780376 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.780385 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.780397 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.780405 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.883338 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.883386 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.883396 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.883411 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.883421 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.986023 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.986104 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.986130 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.986167 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:32 crc kubenswrapper[4935]: I1005 06:54:32.986191 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:32Z","lastTransitionTime":"2025-10-05T06:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.089037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.089100 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.089117 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.089144 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.089160 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.192073 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.192119 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.192135 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.192155 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.192170 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.302771 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.302860 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.302883 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.302939 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.302960 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.406661 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.406708 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.406721 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.406738 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.406750 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.510572 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.510647 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.510667 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.510697 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.510719 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.614201 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.614254 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.614266 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.614284 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.614296 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.717392 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.717463 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.717477 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.717505 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.717521 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.821562 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.821638 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.821772 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.821796 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.821832 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.925143 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.925230 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.925250 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.925281 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:33 crc kubenswrapper[4935]: I1005 06:54:33.925309 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:33Z","lastTransitionTime":"2025-10-05T06:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.027972 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.028092 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.028120 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.028383 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.028417 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.131713 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.131788 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.131805 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.131837 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.131862 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.234966 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.235061 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.235083 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.235112 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.235132 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.338856 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.338963 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.338987 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.339014 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.339032 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.442485 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.442566 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.442594 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.442636 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.442665 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.546513 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.546579 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.546597 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.546687 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.546706 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.650636 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.650723 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.650750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.650782 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.650805 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.753957 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.754004 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.754014 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.754029 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.754042 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.776587 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.776665 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.776707 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.777011 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:34 crc kubenswrapper[4935]: E1005 06:54:34.776965 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:34 crc kubenswrapper[4935]: E1005 06:54:34.777125 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:34 crc kubenswrapper[4935]: E1005 06:54:34.777189 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:34 crc kubenswrapper[4935]: E1005 06:54:34.777445 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.858227 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.858303 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.858360 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.858388 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.858408 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.962077 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.962151 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.962176 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.962213 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:34 crc kubenswrapper[4935]: I1005 06:54:34.962238 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:34Z","lastTransitionTime":"2025-10-05T06:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.065123 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.065188 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.065206 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.065234 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.065600 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.170417 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.170570 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.170591 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.170617 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.170635 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.274079 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.274140 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.274157 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.274182 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.274199 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.378033 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.378090 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.378106 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.378131 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.378150 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.481371 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.481434 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.481452 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.481481 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.481498 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.584011 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.584078 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.584104 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.584128 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.584145 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.690563 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.690613 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.690624 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.690640 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.690653 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.794303 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.794347 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.794361 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.794378 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.794391 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.896691 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.896727 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.896736 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.896750 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.896760 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.999706 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:35 crc kubenswrapper[4935]: I1005 06:54:35.999810 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:35.999933 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:35.999964 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:35.999981 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:35Z","lastTransitionTime":"2025-10-05T06:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.102495 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.102599 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.102620 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.102650 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.102670 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.205712 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.205784 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.205807 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.205831 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.205849 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.309124 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.309583 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.309784 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.309981 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.310146 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.412960 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.413016 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.413037 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.413062 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.413078 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.516196 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.516728 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.517126 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.517322 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.517564 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.620797 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.621356 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.621523 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.621703 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.621918 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.724731 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.724796 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.724808 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.724825 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.724839 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.776404 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:36 crc kubenswrapper[4935]: E1005 06:54:36.776528 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.776600 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.776739 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.776788 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:36 crc kubenswrapper[4935]: E1005 06:54:36.776957 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:36 crc kubenswrapper[4935]: E1005 06:54:36.777079 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:36 crc kubenswrapper[4935]: E1005 06:54:36.777313 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.799223 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=32.799202596 podStartE2EDuration="32.799202596s" podCreationTimestamp="2025-10-05 06:54:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:36.799067862 +0000 UTC m=+110.681694332" watchObservedRunningTime="2025-10-05 06:54:36.799202596 +0000 UTC m=+110.681829096" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.828564 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.828674 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.828695 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.828718 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.828735 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.842300 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=86.842280593 podStartE2EDuration="1m26.842280593s" podCreationTimestamp="2025-10-05 06:53:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:36.822566369 +0000 UTC m=+110.705192899" watchObservedRunningTime="2025-10-05 06:54:36.842280593 +0000 UTC m=+110.724907093" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.844115 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-cz2rj" podStartSLOduration=90.844062271 podStartE2EDuration="1m30.844062271s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:36.843504686 +0000 UTC m=+110.726131186" watchObservedRunningTime="2025-10-05 06:54:36.844062271 +0000 UTC m=+110.726688771" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.858042 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.858016339 podStartE2EDuration="56.858016339s" podCreationTimestamp="2025-10-05 06:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:36.856787636 +0000 UTC m=+110.739414136" watchObservedRunningTime="2025-10-05 06:54:36.858016339 +0000 UTC m=+110.740642829" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.926088 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=12.926071763 podStartE2EDuration="12.926071763s" podCreationTimestamp="2025-10-05 06:54:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:36.92337699 +0000 UTC m=+110.806003450" watchObservedRunningTime="2025-10-05 06:54:36.926071763 +0000 UTC m=+110.808698223" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.930921 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.930960 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.930971 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.930986 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.930999 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:36Z","lastTransitionTime":"2025-10-05T06:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:36 crc kubenswrapper[4935]: I1005 06:54:36.981621 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-w5h5q" podStartSLOduration=90.981602707 podStartE2EDuration="1m30.981602707s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:36.96251975 +0000 UTC m=+110.845146210" watchObservedRunningTime="2025-10-05 06:54:36.981602707 +0000 UTC m=+110.864229167" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.000806 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podStartSLOduration=91.000780887 podStartE2EDuration="1m31.000780887s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:37.000433178 +0000 UTC m=+110.883059658" watchObservedRunningTime="2025-10-05 06:54:37.000780887 +0000 UTC m=+110.883407367" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.018187 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-687xm" podStartSLOduration=91.018168708 podStartE2EDuration="1m31.018168708s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:37.017679735 +0000 UTC m=+110.900306205" watchObservedRunningTime="2025-10-05 06:54:37.018168708 +0000 UTC m=+110.900795168" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.033438 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.033547 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.033556 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.033569 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.033577 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.055524 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=91.0555022 podStartE2EDuration="1m31.0555022s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:37.054576055 +0000 UTC m=+110.937202515" watchObservedRunningTime="2025-10-05 06:54:37.0555022 +0000 UTC m=+110.938128670" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.065510 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-gth9x" podStartSLOduration=91.06549122 podStartE2EDuration="1m31.06549122s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:37.065178552 +0000 UTC m=+110.947805022" watchObservedRunningTime="2025-10-05 06:54:37.06549122 +0000 UTC m=+110.948117680" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.107449 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-znjdd" podStartSLOduration=91.107430407 podStartE2EDuration="1m31.107430407s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:37.10718654 +0000 UTC m=+110.989813020" watchObservedRunningTime="2025-10-05 06:54:37.107430407 +0000 UTC m=+110.990056867" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.136176 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.136231 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.136246 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.136270 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.136285 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.239169 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.239399 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.239471 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.239538 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.239601 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.342864 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.343151 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.343227 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.343297 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.343354 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.445845 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.445923 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.445945 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.445970 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.445989 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.548865 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.548928 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.548947 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.548970 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.548983 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.652665 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.652705 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.652716 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.652732 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.652743 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.755423 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.755501 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.755527 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.755560 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.755584 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.778176 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 06:54:37 crc kubenswrapper[4935]: E1005 06:54:37.778352 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ns4c5_openshift-ovn-kubernetes(019e7dde-1935-4b4c-8312-99f52b6d22fc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.858271 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.858357 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.858385 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.858415 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.858438 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.962654 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.962738 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.962764 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.962796 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:37 crc kubenswrapper[4935]: I1005 06:54:37.962820 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:37Z","lastTransitionTime":"2025-10-05T06:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.069639 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.069699 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.069711 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.069731 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.069743 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.172815 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.172886 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.172954 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.172986 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.173008 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.276922 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.276960 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.276972 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.276991 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.277003 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.379299 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.379351 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.379369 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.379392 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.379409 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.482305 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.482373 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.482390 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.482417 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.482435 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.586108 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.586158 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.586170 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.586188 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.586200 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.689854 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.689939 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.689958 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.689985 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.690003 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.777220 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.777331 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:38 crc kubenswrapper[4935]: E1005 06:54:38.777446 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.777475 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.777230 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:38 crc kubenswrapper[4935]: E1005 06:54:38.777632 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:38 crc kubenswrapper[4935]: E1005 06:54:38.777856 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:38 crc kubenswrapper[4935]: E1005 06:54:38.777962 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.792764 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.792837 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.792862 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.792894 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.792942 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.894813 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.894899 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.894969 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.895000 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.895023 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.898783 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.898832 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.898844 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.898862 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.898874 4935 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:54:38Z","lastTransitionTime":"2025-10-05T06:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.957267 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh"] Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.958546 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.961329 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.965314 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.965365 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 05 06:54:38 crc kubenswrapper[4935]: I1005 06:54:38.965637 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.036227 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.036371 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.036491 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.036587 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.036652 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.137522 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.137601 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.137628 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.137646 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.137671 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.137771 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.137861 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.138577 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.151027 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.170905 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cf7d8dc-5a79-4c8d-a919-69c0934aacd3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hshdh\" (UID: \"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.280498 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" Oct 05 06:54:39 crc kubenswrapper[4935]: W1005 06:54:39.295929 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cf7d8dc_5a79_4c8d_a919_69c0934aacd3.slice/crio-2a1c42fc42e79a7334c5c7c872513169b52e6d337573df4dfec8829767a15807 WatchSource:0}: Error finding container 2a1c42fc42e79a7334c5c7c872513169b52e6d337573df4dfec8829767a15807: Status 404 returned error can't find the container with id 2a1c42fc42e79a7334c5c7c872513169b52e6d337573df4dfec8829767a15807 Oct 05 06:54:39 crc kubenswrapper[4935]: I1005 06:54:39.421139 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" event={"ID":"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3","Type":"ContainerStarted","Data":"2a1c42fc42e79a7334c5c7c872513169b52e6d337573df4dfec8829767a15807"} Oct 05 06:54:40 crc kubenswrapper[4935]: I1005 06:54:40.426864 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" event={"ID":"0cf7d8dc-5a79-4c8d-a919-69c0934aacd3","Type":"ContainerStarted","Data":"69a90031218b7d0be403277d5b7cfbc00c211b2342de5f516922a85a11b2b806"} Oct 05 06:54:40 crc kubenswrapper[4935]: I1005 06:54:40.449441 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hshdh" podStartSLOduration=94.449426292 podStartE2EDuration="1m34.449426292s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:40.44861958 +0000 UTC m=+114.331246080" watchObservedRunningTime="2025-10-05 06:54:40.449426292 +0000 UTC m=+114.332052752" Oct 05 06:54:40 crc kubenswrapper[4935]: I1005 06:54:40.777020 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:40 crc kubenswrapper[4935]: I1005 06:54:40.777059 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:40 crc kubenswrapper[4935]: I1005 06:54:40.777074 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:40 crc kubenswrapper[4935]: I1005 06:54:40.777166 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:40 crc kubenswrapper[4935]: E1005 06:54:40.777172 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:40 crc kubenswrapper[4935]: E1005 06:54:40.777294 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:40 crc kubenswrapper[4935]: E1005 06:54:40.777502 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:40 crc kubenswrapper[4935]: E1005 06:54:40.777694 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:42 crc kubenswrapper[4935]: I1005 06:54:42.776880 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:42 crc kubenswrapper[4935]: E1005 06:54:42.777030 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:42 crc kubenswrapper[4935]: I1005 06:54:42.776877 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:42 crc kubenswrapper[4935]: I1005 06:54:42.777206 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:42 crc kubenswrapper[4935]: I1005 06:54:42.777289 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:42 crc kubenswrapper[4935]: E1005 06:54:42.777272 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:42 crc kubenswrapper[4935]: E1005 06:54:42.777361 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:42 crc kubenswrapper[4935]: E1005 06:54:42.777496 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:43 crc kubenswrapper[4935]: I1005 06:54:43.438216 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/1.log" Oct 05 06:54:43 crc kubenswrapper[4935]: I1005 06:54:43.439149 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/0.log" Oct 05 06:54:43 crc kubenswrapper[4935]: I1005 06:54:43.439274 4935 generic.go:334] "Generic (PLEG): container finished" podID="2b9ff491-4f10-4f1b-822b-a8c46871ce60" containerID="7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79" exitCode=1 Oct 05 06:54:43 crc kubenswrapper[4935]: I1005 06:54:43.439337 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-687xm" event={"ID":"2b9ff491-4f10-4f1b-822b-a8c46871ce60","Type":"ContainerDied","Data":"7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79"} Oct 05 06:54:43 crc kubenswrapper[4935]: I1005 06:54:43.439434 4935 scope.go:117] "RemoveContainer" containerID="7d5e1f577dd51df3f0e600709b56f710e3630d68d4f36a31d4808b7dab718dea" Oct 05 06:54:43 crc kubenswrapper[4935]: I1005 06:54:43.440015 4935 scope.go:117] "RemoveContainer" containerID="7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79" Oct 05 06:54:43 crc kubenswrapper[4935]: E1005 06:54:43.440422 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-687xm_openshift-multus(2b9ff491-4f10-4f1b-822b-a8c46871ce60)\"" pod="openshift-multus/multus-687xm" podUID="2b9ff491-4f10-4f1b-822b-a8c46871ce60" Oct 05 06:54:44 crc kubenswrapper[4935]: I1005 06:54:44.445086 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/1.log" Oct 05 06:54:44 crc kubenswrapper[4935]: I1005 06:54:44.777222 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:44 crc kubenswrapper[4935]: I1005 06:54:44.777292 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:44 crc kubenswrapper[4935]: I1005 06:54:44.777234 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:44 crc kubenswrapper[4935]: I1005 06:54:44.777454 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:44 crc kubenswrapper[4935]: E1005 06:54:44.777453 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:44 crc kubenswrapper[4935]: E1005 06:54:44.777623 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:44 crc kubenswrapper[4935]: E1005 06:54:44.777721 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:44 crc kubenswrapper[4935]: E1005 06:54:44.777880 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:46 crc kubenswrapper[4935]: E1005 06:54:46.769091 4935 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 05 06:54:46 crc kubenswrapper[4935]: I1005 06:54:46.776224 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:46 crc kubenswrapper[4935]: E1005 06:54:46.777454 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:46 crc kubenswrapper[4935]: I1005 06:54:46.777533 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:46 crc kubenswrapper[4935]: I1005 06:54:46.777545 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:46 crc kubenswrapper[4935]: E1005 06:54:46.777674 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:46 crc kubenswrapper[4935]: I1005 06:54:46.777857 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:46 crc kubenswrapper[4935]: E1005 06:54:46.777982 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:46 crc kubenswrapper[4935]: E1005 06:54:46.778404 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:46 crc kubenswrapper[4935]: E1005 06:54:46.920146 4935 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 06:54:48 crc kubenswrapper[4935]: I1005 06:54:48.776409 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:48 crc kubenswrapper[4935]: I1005 06:54:48.776418 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:48 crc kubenswrapper[4935]: I1005 06:54:48.776480 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:48 crc kubenswrapper[4935]: I1005 06:54:48.777612 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:48 crc kubenswrapper[4935]: E1005 06:54:48.777731 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:48 crc kubenswrapper[4935]: E1005 06:54:48.777823 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:48 crc kubenswrapper[4935]: E1005 06:54:48.777944 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:48 crc kubenswrapper[4935]: E1005 06:54:48.778128 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:48 crc kubenswrapper[4935]: I1005 06:54:48.778203 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 06:54:49 crc kubenswrapper[4935]: I1005 06:54:49.464224 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/3.log" Oct 05 06:54:49 crc kubenswrapper[4935]: I1005 06:54:49.467556 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerStarted","Data":"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc"} Oct 05 06:54:49 crc kubenswrapper[4935]: I1005 06:54:49.468062 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:54:49 crc kubenswrapper[4935]: I1005 06:54:49.505279 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podStartSLOduration=103.505259213 podStartE2EDuration="1m43.505259213s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:54:49.503246049 +0000 UTC m=+123.385872519" watchObservedRunningTime="2025-10-05 06:54:49.505259213 +0000 UTC m=+123.387885673" Oct 05 06:54:49 crc kubenswrapper[4935]: I1005 06:54:49.636780 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-85tzm"] Oct 05 06:54:49 crc kubenswrapper[4935]: I1005 06:54:49.636958 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:49 crc kubenswrapper[4935]: E1005 06:54:49.637104 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:50 crc kubenswrapper[4935]: I1005 06:54:50.776615 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:50 crc kubenswrapper[4935]: I1005 06:54:50.776620 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:50 crc kubenswrapper[4935]: E1005 06:54:50.777298 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:50 crc kubenswrapper[4935]: E1005 06:54:50.777388 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:50 crc kubenswrapper[4935]: I1005 06:54:50.776676 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:50 crc kubenswrapper[4935]: E1005 06:54:50.777735 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:51 crc kubenswrapper[4935]: I1005 06:54:51.776433 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:51 crc kubenswrapper[4935]: E1005 06:54:51.776641 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:51 crc kubenswrapper[4935]: E1005 06:54:51.921919 4935 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 06:54:52 crc kubenswrapper[4935]: I1005 06:54:52.777176 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:52 crc kubenswrapper[4935]: I1005 06:54:52.777214 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:52 crc kubenswrapper[4935]: I1005 06:54:52.777283 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:52 crc kubenswrapper[4935]: E1005 06:54:52.777361 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:52 crc kubenswrapper[4935]: E1005 06:54:52.777477 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:52 crc kubenswrapper[4935]: E1005 06:54:52.777601 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:53 crc kubenswrapper[4935]: I1005 06:54:53.777134 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:53 crc kubenswrapper[4935]: E1005 06:54:53.777327 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:54 crc kubenswrapper[4935]: I1005 06:54:54.777188 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:54 crc kubenswrapper[4935]: E1005 06:54:54.777341 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:54 crc kubenswrapper[4935]: I1005 06:54:54.777463 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:54 crc kubenswrapper[4935]: I1005 06:54:54.778088 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:54 crc kubenswrapper[4935]: E1005 06:54:54.778165 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:54 crc kubenswrapper[4935]: E1005 06:54:54.778283 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:55 crc kubenswrapper[4935]: I1005 06:54:55.776508 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:55 crc kubenswrapper[4935]: E1005 06:54:55.776658 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:56 crc kubenswrapper[4935]: I1005 06:54:56.776774 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:56 crc kubenswrapper[4935]: E1005 06:54:56.778170 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:56 crc kubenswrapper[4935]: I1005 06:54:56.778272 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:56 crc kubenswrapper[4935]: I1005 06:54:56.778356 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:56 crc kubenswrapper[4935]: E1005 06:54:56.778714 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:56 crc kubenswrapper[4935]: E1005 06:54:56.778830 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:56 crc kubenswrapper[4935]: I1005 06:54:56.778973 4935 scope.go:117] "RemoveContainer" containerID="7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79" Oct 05 06:54:56 crc kubenswrapper[4935]: E1005 06:54:56.923619 4935 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 06:54:57 crc kubenswrapper[4935]: I1005 06:54:57.498414 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/1.log" Oct 05 06:54:57 crc kubenswrapper[4935]: I1005 06:54:57.498687 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-687xm" event={"ID":"2b9ff491-4f10-4f1b-822b-a8c46871ce60","Type":"ContainerStarted","Data":"d22b4290c09db56259541f5d949e9b59e16b541d40ce02e695c7ba63bcb3961f"} Oct 05 06:54:57 crc kubenswrapper[4935]: I1005 06:54:57.776591 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:57 crc kubenswrapper[4935]: E1005 06:54:57.777040 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:54:58 crc kubenswrapper[4935]: I1005 06:54:58.776366 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:54:58 crc kubenswrapper[4935]: I1005 06:54:58.776464 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:54:58 crc kubenswrapper[4935]: I1005 06:54:58.776536 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:54:58 crc kubenswrapper[4935]: E1005 06:54:58.776530 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:54:58 crc kubenswrapper[4935]: E1005 06:54:58.776673 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:54:58 crc kubenswrapper[4935]: E1005 06:54:58.776810 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:54:59 crc kubenswrapper[4935]: I1005 06:54:59.776833 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:54:59 crc kubenswrapper[4935]: E1005 06:54:59.777013 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:55:00 crc kubenswrapper[4935]: I1005 06:55:00.776536 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:55:00 crc kubenswrapper[4935]: I1005 06:55:00.776619 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:55:00 crc kubenswrapper[4935]: I1005 06:55:00.776545 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:55:00 crc kubenswrapper[4935]: E1005 06:55:00.776733 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:55:00 crc kubenswrapper[4935]: E1005 06:55:00.776792 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:55:00 crc kubenswrapper[4935]: E1005 06:55:00.776850 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:55:01 crc kubenswrapper[4935]: I1005 06:55:01.776810 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:55:01 crc kubenswrapper[4935]: E1005 06:55:01.777152 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-85tzm" podUID="87a0c84e-d575-4468-af55-c814e305d34d" Oct 05 06:55:02 crc kubenswrapper[4935]: I1005 06:55:02.776961 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:55:02 crc kubenswrapper[4935]: I1005 06:55:02.777055 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:55:02 crc kubenswrapper[4935]: I1005 06:55:02.777559 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:55:02 crc kubenswrapper[4935]: I1005 06:55:02.780954 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 05 06:55:02 crc kubenswrapper[4935]: I1005 06:55:02.781012 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 05 06:55:02 crc kubenswrapper[4935]: I1005 06:55:02.781224 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 05 06:55:02 crc kubenswrapper[4935]: I1005 06:55:02.781335 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 05 06:55:03 crc kubenswrapper[4935]: I1005 06:55:03.776596 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:55:03 crc kubenswrapper[4935]: I1005 06:55:03.779328 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 05 06:55:03 crc kubenswrapper[4935]: I1005 06:55:03.780469 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 05 06:55:06 crc kubenswrapper[4935]: I1005 06:55:06.992289 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.485561 4935 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.542799 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rphcb"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.543588 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.547234 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.548491 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.551558 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.552551 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f254r"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.556348 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.557363 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.560186 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.560418 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.565048 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.565130 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.565604 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-44xpc"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.565983 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.566120 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.566753 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.566904 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.568172 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.568445 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.568761 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.568795 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.569040 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.570761 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.571536 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.572420 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.572699 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.573064 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.573115 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.573436 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.577957 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nh99w"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.579001 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.580829 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4064a1c6-00f1-4c14-9ce9-42308f63069f-config\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.580910 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8d2c\" (UniqueName: \"kubernetes.io/projected/4064a1c6-00f1-4c14-9ce9-42308f63069f-kube-api-access-x8d2c\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.580980 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4064a1c6-00f1-4c14-9ce9-42308f63069f-images\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.581011 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4064a1c6-00f1-4c14-9ce9-42308f63069f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.586986 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.587817 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.595477 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.595822 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.596348 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.596834 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.596970 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.597300 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.597577 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.598198 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.598592 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.598636 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.598732 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.598862 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599000 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599136 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599172 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599170 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599230 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599253 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599149 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599442 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599466 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.599662 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.600623 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.601207 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.601648 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.601938 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.630301 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9hwvw"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.634770 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.643081 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.643463 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cn4mw"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.643166 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.643927 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.646869 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.647328 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.647736 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.648014 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.648314 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.648541 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.651083 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.651684 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.654364 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.655102 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-cbb4c"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.655682 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.656135 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.656529 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.656869 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.657088 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.659267 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.659444 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.660053 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.660221 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.660340 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.662055 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.662072 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.663258 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.670263 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6tkbm"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.670642 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.670941 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.672111 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.672119 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.673532 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.674468 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.674922 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.675078 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.675781 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-trqss"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.675994 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.676570 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.677023 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.677582 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.678267 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.678582 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.678858 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.678917 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.679612 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.679860 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.680061 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.680169 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.680265 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.680927 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.682021 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.682130 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.682337 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.682363 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-client-ca\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.682500 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8d2c\" (UniqueName: \"kubernetes.io/projected/4064a1c6-00f1-4c14-9ce9-42308f63069f-kube-api-access-x8d2c\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.682525 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fb78a87-5385-4785-b4b4-4e077d5000ed-serving-cert\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.705791 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hwbx\" (UniqueName: \"kubernetes.io/projected/e7e163b1-79e1-4547-b3c9-4a506b887337-kube-api-access-5hwbx\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706006 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4064a1c6-00f1-4c14-9ce9-42308f63069f-config\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706038 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-trusted-ca-bundle\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706252 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706462 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706503 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706708 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-console-config\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706743 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-etcd-serving-ca\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685344 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706936 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7039dd6e-0052-4ea2-8a08-be11c457726d-audit-dir\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685401 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.707100 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.707148 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66lb8\" (UniqueName: \"kubernetes.io/projected/7039dd6e-0052-4ea2-8a08-be11c457726d-kube-api-access-66lb8\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.707184 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-client-ca\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.707204 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3ccb7e54-6381-4dfd-92c8-679d33356e40-auth-proxy-config\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.707415 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.707531 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.708085 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4064a1c6-00f1-4c14-9ce9-42308f63069f-config\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.707613 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-785ds\" (UniqueName: \"kubernetes.io/projected/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-kube-api-access-785ds\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.708335 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-audit\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.708425 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.708532 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-config\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.708559 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkmh6\" (UniqueName: \"kubernetes.io/projected/8088d77f-c1e1-4469-af37-304d63c82f1b-kube-api-access-dkmh6\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.709462 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e163b1-79e1-4547-b3c9-4a506b887337-serving-cert\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.709492 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c167120c-3675-4125-a194-8a2d22959439-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685447 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685459 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.709693 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rr7r\" (UniqueName: \"kubernetes.io/projected/d90f77af-0ce4-4b2f-87d2-049e45327404-kube-api-access-8rr7r\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.709716 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hcpg\" (UniqueName: \"kubernetes.io/projected/3ccb7e54-6381-4dfd-92c8-679d33356e40-kube-api-access-5hcpg\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685582 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685636 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685669 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685702 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.711005 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.685766 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.711763 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.686142 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.706453 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.712083 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.712151 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sbwkd"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.712341 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.712568 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.714121 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.714383 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.714684 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c167120c-3675-4125-a194-8a2d22959439-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.714986 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hgg5\" (UniqueName: \"kubernetes.io/projected/ac0661f5-9317-459e-bfce-c2ff9058c319-kube-api-access-4hgg5\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.715016 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nxrl\" (UniqueName: \"kubernetes.io/projected/14973180-053d-4f7b-9f72-7acc0f8b19ac-kube-api-access-9nxrl\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.715044 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-service-ca\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.715067 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-serving-cert\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.715089 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.715116 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e163b1-79e1-4547-b3c9-4a506b887337-config\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.715139 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c2923579-4c95-4150-9786-878a133cc0a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zglwh\" (UID: \"c2923579-4c95-4150-9786-878a133cc0a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.718205 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-serving-cert\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.718300 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6kws\" (UniqueName: \"kubernetes.io/projected/c167120c-3675-4125-a194-8a2d22959439-kube-api-access-p6kws\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.718350 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-etcd-client\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.718381 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-service-ca-bundle\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.719477 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.723620 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.724214 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-oauth-config\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.726819 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-node-pullsecrets\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.726983 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727070 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-config\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727154 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727231 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-encryption-config\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727301 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3ccb7e54-6381-4dfd-92c8-679d33356e40-machine-approver-tls\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727385 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7e163b1-79e1-4547-b3c9-4a506b887337-trusted-ca\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727456 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ccb7e54-6381-4dfd-92c8-679d33356e40-config\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727535 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-audit-dir\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727607 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727679 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dsr4\" (UniqueName: \"kubernetes.io/projected/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-kube-api-access-7dsr4\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727754 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727828 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-etcd-client\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727915 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-audit-policies\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.727994 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.728079 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d90f77af-0ce4-4b2f-87d2-049e45327404-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.728190 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4064a1c6-00f1-4c14-9ce9-42308f63069f-images\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.728323 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4064a1c6-00f1-4c14-9ce9-42308f63069f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.728731 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.728830 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j2nz\" (UniqueName: \"kubernetes.io/projected/c2923579-4c95-4150-9786-878a133cc0a4-kube-api-access-6j2nz\") pod \"cluster-samples-operator-665b6dd947-zglwh\" (UID: \"c2923579-4c95-4150-9786-878a133cc0a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.728975 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-image-import-ca\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.729094 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-config\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.729196 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-dir\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.729317 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-oauth-serving-cert\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.729603 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-encryption-config\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.729710 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-config\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.729811 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-serving-cert\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.729910 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.730014 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.730143 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.730225 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bgr8\" (UniqueName: \"kubernetes.io/projected/4fb78a87-5385-4785-b4b4-4e077d5000ed-kube-api-access-9bgr8\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.730311 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8088d77f-c1e1-4469-af37-304d63c82f1b-serving-cert\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.730570 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.730657 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.730740 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-policies\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.731035 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.729845 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.731978 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4064a1c6-00f1-4c14-9ce9-42308f63069f-images\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.733175 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-597nr"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.733400 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.736026 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.736393 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.736551 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-597nr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.736840 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.737440 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mzckv"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.737598 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.737766 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.737932 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.738161 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.738250 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.738312 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.738548 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.738908 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.738975 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.739504 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.739970 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.740576 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.741118 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5skck"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.742251 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.743871 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-gc2gr"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.743968 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4064a1c6-00f1-4c14-9ce9-42308f63069f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.744411 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.746991 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.747616 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.747994 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.748016 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.748061 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.747994 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.748346 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.749436 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.749508 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.749911 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.752083 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.753094 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.754035 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.757091 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-drgxn"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.757814 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.758876 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jbxvd"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.759719 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.761361 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.770960 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.806604 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.806851 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvn45"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.806940 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.807900 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.822800 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.823631 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-68swt"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833308 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833559 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-audit-policies\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833691 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833817 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833858 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23018ad9-0e58-4976-a215-fae5b1b51da6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833907 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j2nz\" (UniqueName: \"kubernetes.io/projected/c2923579-4c95-4150-9786-878a133cc0a4-kube-api-access-6j2nz\") pod \"cluster-samples-operator-665b6dd947-zglwh\" (UID: \"c2923579-4c95-4150-9786-878a133cc0a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833929 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qh47\" (UniqueName: \"kubernetes.io/projected/8df3f651-d6cd-48a2-a97e-84448ce9fb00-kube-api-access-8qh47\") pod \"package-server-manager-789f6589d5-r798b\" (UID: \"8df3f651-d6cd-48a2-a97e-84448ce9fb00\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833972 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-image-import-ca\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.833989 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4a6790bc-0770-4667-94d9-4cc203809743-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6tkbm\" (UID: \"4a6790bc-0770-4667-94d9-4cc203809743\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834010 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnjdm\" (UniqueName: \"kubernetes.io/projected/ed54eb3b-78b4-4c57-bb72-49d830aede0d-kube-api-access-hnjdm\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834052 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-dir\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834074 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c26rq\" (UniqueName: \"kubernetes.io/projected/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-kube-api-access-c26rq\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834135 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zfm6\" (UniqueName: \"kubernetes.io/projected/c33e835e-eefc-4c71-84ba-a7e6784b17bd-kube-api-access-9zfm6\") pod \"control-plane-machine-set-operator-78cbb6b69f-zp4tc\" (UID: \"c33e835e-eefc-4c71-84ba-a7e6784b17bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834154 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0a9c01-eaff-4da6-a599-48af41b70260-serving-cert\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834176 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-config\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834211 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834251 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsn2q\" (UniqueName: \"kubernetes.io/projected/22367e16-ef10-498d-8cad-b719508ea6eb-kube-api-access-lsn2q\") pod \"downloads-7954f5f757-597nr\" (UID: \"22367e16-ef10-498d-8cad-b719508ea6eb\") " pod="openshift-console/downloads-7954f5f757-597nr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834271 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ef91e14-6bf7-47de-b6da-ec630e189389-service-ca-bundle\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834290 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834463 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bgr8\" (UniqueName: \"kubernetes.io/projected/4fb78a87-5385-4785-b4b4-4e077d5000ed-kube-api-access-9bgr8\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834561 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-audit-policies\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.834494 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.835494 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-image-import-ca\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.835584 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-dir\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.836435 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.836559 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837292 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837077 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837375 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhvn2\" (UniqueName: \"kubernetes.io/projected/4a6790bc-0770-4667-94d9-4cc203809743-kube-api-access-nhvn2\") pod \"multus-admission-controller-857f4d67dd-6tkbm\" (UID: \"4a6790bc-0770-4667-94d9-4cc203809743\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837404 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4shb7\" (UniqueName: \"kubernetes.io/projected/ac0a9c01-eaff-4da6-a599-48af41b70260-kube-api-access-4shb7\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837437 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-apiservice-cert\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837463 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-default-certificate\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837484 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-secret-volume\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837502 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c55e2cd9-3e55-492c-b005-d30c278bfb03-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837588 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837615 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837638 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-console-config\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837659 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66lb8\" (UniqueName: \"kubernetes.io/projected/7039dd6e-0052-4ea2-8a08-be11c457726d-kube-api-access-66lb8\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-config\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837681 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-785ds\" (UniqueName: \"kubernetes.io/projected/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-kube-api-access-785ds\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837703 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-client-ca\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837660 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837724 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3ccb7e54-6381-4dfd-92c8-679d33356e40-auth-proxy-config\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837804 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkmh6\" (UniqueName: \"kubernetes.io/projected/8088d77f-c1e1-4469-af37-304d63c82f1b-kube-api-access-dkmh6\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837846 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rr7r\" (UniqueName: \"kubernetes.io/projected/d90f77af-0ce4-4b2f-87d2-049e45327404-kube-api-access-8rr7r\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837887 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0bae24b-7c92-4d48-a608-9b7a62c37161-serving-cert\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837944 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vj85\" (UniqueName: \"kubernetes.io/projected/c399d3f0-061d-49b7-beb8-3bc3dc378b39-kube-api-access-9vj85\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.837975 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlv5m\" (UniqueName: \"kubernetes.io/projected/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-kube-api-access-jlv5m\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838012 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hgg5\" (UniqueName: \"kubernetes.io/projected/ac0661f5-9317-459e-bfce-c2ff9058c319-kube-api-access-4hgg5\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838043 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nxrl\" (UniqueName: \"kubernetes.io/projected/14973180-053d-4f7b-9f72-7acc0f8b19ac-kube-api-access-9nxrl\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838067 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxpzv\" (UniqueName: \"kubernetes.io/projected/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-kube-api-access-kxpzv\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838086 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-proxy-tls\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838110 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838398 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23018ad9-0e58-4976-a215-fae5b1b51da6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838506 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-service-ca\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838599 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-serving-cert\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838683 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-etcd-client\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838769 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmssh\" (UniqueName: \"kubernetes.io/projected/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-kube-api-access-mmssh\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838856 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-oauth-config\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839131 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3ccb7e54-6381-4dfd-92c8-679d33356e40-auth-proxy-config\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.838695 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-console-config\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839145 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-config\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839518 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ddea8a8-fd20-4eb2-8321-98f52f5847da-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839614 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23018ad9-0e58-4976-a215-fae5b1b51da6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839672 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ddea8a8-fd20-4eb2-8321-98f52f5847da-proxy-tls\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839802 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839845 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgnp9\" (UniqueName: \"kubernetes.io/projected/8f91ddae-6adc-49a0-ac67-dd37606d3427-kube-api-access-dgnp9\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839925 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839963 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.839937 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-metrics-certs\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840266 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-68swt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840423 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-service-ca\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840566 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-config\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840614 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ccb7e54-6381-4dfd-92c8-679d33356e40-config\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840654 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-service-ca\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840679 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840741 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dsr4\" (UniqueName: \"kubernetes.io/projected/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-kube-api-access-7dsr4\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840762 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d90f77af-0ce4-4b2f-87d2-049e45327404-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840783 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-ca\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840804 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-client\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840858 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cc4744df-9673-48f7-86cb-51df34fcaf17-metrics-tls\") pod \"dns-operator-744455d44c-sbwkd\" (UID: \"cc4744df-9673-48f7-86cb-51df34fcaf17\") " pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840902 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840929 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed54eb3b-78b4-4c57-bb72-49d830aede0d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840955 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-config\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840987 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tffwc\" (UniqueName: \"kubernetes.io/projected/b0bae24b-7c92-4d48-a608-9b7a62c37161-kube-api-access-tffwc\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.841009 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-images\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.840914 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.841450 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.841826 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.841924 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-config\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.842016 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.842377 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-config\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.842397 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ccb7e54-6381-4dfd-92c8-679d33356e40-config\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.842416 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-oauth-serving-cert\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.842442 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-encryption-config\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.842974 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-etcd-client\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.843553 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-client-ca\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.843819 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-srv-cert\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.843850 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-serving-cert\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.843913 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-profile-collector-cert\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.843938 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.844244 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.844306 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-44xpc"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.845086 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-oauth-serving-cert\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.845195 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.845255 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8088d77f-c1e1-4469-af37-304d63c82f1b-serving-cert\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.845279 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-policies\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.845343 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c399d3f0-061d-49b7-beb8-3bc3dc378b39-tmpfs\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.845821 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-policies\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.846446 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.846567 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-config\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.846741 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.846969 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbpg2\" (UniqueName: \"kubernetes.io/projected/671f9210-cd88-458e-aec5-eceaadcc3f23-kube-api-access-zbpg2\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.847061 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.847143 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-client-ca\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.847266 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fb78a87-5385-4785-b4b4-4e077d5000ed-serving-cert\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.847334 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hwbx\" (UniqueName: \"kubernetes.io/projected/e7e163b1-79e1-4547-b3c9-4a506b887337-kube-api-access-5hwbx\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.847571 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d90f77af-0ce4-4b2f-87d2-049e45327404-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.847810 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-oauth-config\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.850600 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-serving-cert\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.850917 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-encryption-config\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851160 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851284 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-serving-cert\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851457 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-client-ca\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851699 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-trusted-ca-bundle\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851749 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851778 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851799 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851819 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c55e2cd9-3e55-492c-b005-d30c278bfb03-config\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851840 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-etcd-serving-ca\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851856 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7039dd6e-0052-4ea2-8a08-be11c457726d-audit-dir\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851873 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851903 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851935 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5np92\" (UniqueName: \"kubernetes.io/projected/cc4744df-9673-48f7-86cb-51df34fcaf17-kube-api-access-5np92\") pod \"dns-operator-744455d44c-sbwkd\" (UID: \"cc4744df-9673-48f7-86cb-51df34fcaf17\") " pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.851985 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t522c\" (UniqueName: \"kubernetes.io/projected/8ef91e14-6bf7-47de-b6da-ec630e189389-kube-api-access-t522c\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852019 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-audit\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852044 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-config\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852073 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e163b1-79e1-4547-b3c9-4a506b887337-serving-cert\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852097 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c167120c-3675-4125-a194-8a2d22959439-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852120 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hcpg\" (UniqueName: \"kubernetes.io/projected/3ccb7e54-6381-4dfd-92c8-679d33356e40-kube-api-access-5hcpg\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852144 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-metrics-tls\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852167 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852160 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cn4mw"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852190 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c167120c-3675-4125-a194-8a2d22959439-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852221 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed54eb3b-78b4-4c57-bb72-49d830aede0d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852250 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852279 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-serving-cert\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852302 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852314 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7039dd6e-0052-4ea2-8a08-be11c457726d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852327 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8df3f651-d6cd-48a2-a97e-84448ce9fb00-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r798b\" (UID: \"8df3f651-d6cd-48a2-a97e-84448ce9fb00\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852354 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e163b1-79e1-4547-b3c9-4a506b887337-config\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852378 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c2923579-4c95-4150-9786-878a133cc0a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zglwh\" (UID: \"c2923579-4c95-4150-9786-878a133cc0a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852435 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-cabundle\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852461 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c55e2cd9-3e55-492c-b005-d30c278bfb03-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852494 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6kws\" (UniqueName: \"kubernetes.io/projected/c167120c-3675-4125-a194-8a2d22959439-kube-api-access-p6kws\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852519 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-key\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852542 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvcw7\" (UniqueName: \"kubernetes.io/projected/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-kube-api-access-kvcw7\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852565 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/671f9210-cd88-458e-aec5-eceaadcc3f23-srv-cert\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852596 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-service-ca-bundle\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852622 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852646 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-webhook-cert\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852668 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac0a9c01-eaff-4da6-a599-48af41b70260-config\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852688 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-trusted-ca\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852722 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-serving-cert\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852750 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-node-pullsecrets\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852774 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852797 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-stats-auth\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852837 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5qbp\" (UniqueName: \"kubernetes.io/projected/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-kube-api-access-j5qbp\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852861 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c33e835e-eefc-4c71-84ba-a7e6784b17bd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zp4tc\" (UID: \"c33e835e-eefc-4c71-84ba-a7e6784b17bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852900 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwwgj\" (UniqueName: \"kubernetes.io/projected/7ddea8a8-fd20-4eb2-8321-98f52f5847da-kube-api-access-gwwgj\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852928 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-encryption-config\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852952 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3ccb7e54-6381-4dfd-92c8-679d33356e40-machine-approver-tls\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852950 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-trusted-ca-bundle\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852973 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-etcd-serving-ca\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852977 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/671f9210-cd88-458e-aec5-eceaadcc3f23-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.853034 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.852355 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7039dd6e-0052-4ea2-8a08-be11c457726d-audit-dir\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.853186 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.853263 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7e163b1-79e1-4547-b3c9-4a506b887337-trusted-ca\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.853300 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-audit-dir\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.853373 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-etcd-client\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.853404 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.853432 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bvx8\" (UniqueName: \"kubernetes.io/projected/c1af3cc3-3fad-473d-a193-7de470b5ff7c-kube-api-access-6bvx8\") pod \"migrator-59844c95c7-7ckxt\" (UID: \"c1af3cc3-3fad-473d-a193-7de470b5ff7c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.853457 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-bound-sa-token\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.855042 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-audit\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.855094 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.855343 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8088d77f-c1e1-4469-af37-304d63c82f1b-serving-cert\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.855798 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fb78a87-5385-4785-b4b4-4e077d5000ed-serving-cert\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.855942 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-service-ca-bundle\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.856533 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c167120c-3675-4125-a194-8a2d22959439-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.856605 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-node-pullsecrets\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.856998 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7039dd6e-0052-4ea2-8a08-be11c457726d-encryption-config\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.857259 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.857708 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.859223 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3ccb7e54-6381-4dfd-92c8-679d33356e40-machine-approver-tls\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.859270 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.859439 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nh99w"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.859556 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-audit-dir\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.859567 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sbwkd"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.859524 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-etcd-client\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.860243 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e163b1-79e1-4547-b3c9-4a506b887337-config\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.860584 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8088d77f-c1e1-4469-af37-304d63c82f1b-config\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.861490 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.861545 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.861716 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c2923579-4c95-4150-9786-878a133cc0a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zglwh\" (UID: \"c2923579-4c95-4150-9786-878a133cc0a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.862037 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-trqss"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.862060 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e163b1-79e1-4547-b3c9-4a506b887337-serving-cert\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.862446 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.862507 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7e163b1-79e1-4547-b3c9-4a506b887337-trusted-ca\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.863121 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.863243 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6tkbm"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.863837 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-serving-cert\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.864064 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c167120c-3675-4125-a194-8a2d22959439-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.864455 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f254r"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.865658 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.865833 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.867125 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9hwvw"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.868160 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5skck"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.869731 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.870528 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.871536 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rphcb"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.872685 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-597nr"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.873620 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-zdg8q"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.874632 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.875015 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.875293 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.878087 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ssjkm"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.879573 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.879591 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-cbb4c"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.879729 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.881656 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.882719 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.884980 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.886036 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.887285 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.887624 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mzckv"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.888498 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.890159 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.893617 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.893675 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvn45"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.894011 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jbxvd"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.896951 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-68swt"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.898284 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.899378 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-drgxn"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.900765 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ssjkm"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.902013 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-bskzq"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.903249 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.903710 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bskzq"] Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.921304 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8d2c\" (UniqueName: \"kubernetes.io/projected/4064a1c6-00f1-4c14-9ce9-42308f63069f-kube-api-access-x8d2c\") pod \"machine-api-operator-5694c8668f-rphcb\" (UID: \"4064a1c6-00f1-4c14-9ce9-42308f63069f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.945751 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954573 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbpg2\" (UniqueName: \"kubernetes.io/projected/671f9210-cd88-458e-aec5-eceaadcc3f23-kube-api-access-zbpg2\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954660 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954697 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c55e2cd9-3e55-492c-b005-d30c278bfb03-config\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954724 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t522c\" (UniqueName: \"kubernetes.io/projected/8ef91e14-6bf7-47de-b6da-ec630e189389-kube-api-access-t522c\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954749 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5np92\" (UniqueName: \"kubernetes.io/projected/cc4744df-9673-48f7-86cb-51df34fcaf17-kube-api-access-5np92\") pod \"dns-operator-744455d44c-sbwkd\" (UID: \"cc4744df-9673-48f7-86cb-51df34fcaf17\") " pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954768 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-metrics-tls\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954799 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed54eb3b-78b4-4c57-bb72-49d830aede0d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954819 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954838 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8df3f651-d6cd-48a2-a97e-84448ce9fb00-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r798b\" (UID: \"8df3f651-d6cd-48a2-a97e-84448ce9fb00\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954861 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-cabundle\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954876 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c55e2cd9-3e55-492c-b005-d30c278bfb03-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.954978 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-key\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955001 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvcw7\" (UniqueName: \"kubernetes.io/projected/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-kube-api-access-kvcw7\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955022 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/671f9210-cd88-458e-aec5-eceaadcc3f23-srv-cert\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955070 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-webhook-cert\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955091 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac0a9c01-eaff-4da6-a599-48af41b70260-config\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955112 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-trusted-ca\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955129 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-serving-cert\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955151 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-stats-auth\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955171 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5qbp\" (UniqueName: \"kubernetes.io/projected/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-kube-api-access-j5qbp\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955194 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c33e835e-eefc-4c71-84ba-a7e6784b17bd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zp4tc\" (UID: \"c33e835e-eefc-4c71-84ba-a7e6784b17bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955219 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwwgj\" (UniqueName: \"kubernetes.io/projected/7ddea8a8-fd20-4eb2-8321-98f52f5847da-kube-api-access-gwwgj\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955240 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/671f9210-cd88-458e-aec5-eceaadcc3f23-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955262 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bvx8\" (UniqueName: \"kubernetes.io/projected/c1af3cc3-3fad-473d-a193-7de470b5ff7c-kube-api-access-6bvx8\") pod \"migrator-59844c95c7-7ckxt\" (UID: \"c1af3cc3-3fad-473d-a193-7de470b5ff7c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955283 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-bound-sa-token\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955317 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955339 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23018ad9-0e58-4976-a215-fae5b1b51da6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955364 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qh47\" (UniqueName: \"kubernetes.io/projected/8df3f651-d6cd-48a2-a97e-84448ce9fb00-kube-api-access-8qh47\") pod \"package-server-manager-789f6589d5-r798b\" (UID: \"8df3f651-d6cd-48a2-a97e-84448ce9fb00\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955384 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4a6790bc-0770-4667-94d9-4cc203809743-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6tkbm\" (UID: \"4a6790bc-0770-4667-94d9-4cc203809743\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955404 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnjdm\" (UniqueName: \"kubernetes.io/projected/ed54eb3b-78b4-4c57-bb72-49d830aede0d-kube-api-access-hnjdm\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955442 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c26rq\" (UniqueName: \"kubernetes.io/projected/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-kube-api-access-c26rq\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955459 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zfm6\" (UniqueName: \"kubernetes.io/projected/c33e835e-eefc-4c71-84ba-a7e6784b17bd-kube-api-access-9zfm6\") pod \"control-plane-machine-set-operator-78cbb6b69f-zp4tc\" (UID: \"c33e835e-eefc-4c71-84ba-a7e6784b17bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955475 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0a9c01-eaff-4da6-a599-48af41b70260-serving-cert\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955493 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsn2q\" (UniqueName: \"kubernetes.io/projected/22367e16-ef10-498d-8cad-b719508ea6eb-kube-api-access-lsn2q\") pod \"downloads-7954f5f757-597nr\" (UID: \"22367e16-ef10-498d-8cad-b719508ea6eb\") " pod="openshift-console/downloads-7954f5f757-597nr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955509 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ef91e14-6bf7-47de-b6da-ec630e189389-service-ca-bundle\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955544 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955574 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhvn2\" (UniqueName: \"kubernetes.io/projected/4a6790bc-0770-4667-94d9-4cc203809743-kube-api-access-nhvn2\") pod \"multus-admission-controller-857f4d67dd-6tkbm\" (UID: \"4a6790bc-0770-4667-94d9-4cc203809743\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955592 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4shb7\" (UniqueName: \"kubernetes.io/projected/ac0a9c01-eaff-4da6-a599-48af41b70260-kube-api-access-4shb7\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955611 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-apiservice-cert\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955626 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-default-certificate\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955641 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-secret-volume\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955670 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c55e2cd9-3e55-492c-b005-d30c278bfb03-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955688 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955736 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0bae24b-7c92-4d48-a608-9b7a62c37161-serving-cert\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.955765 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vj85\" (UniqueName: \"kubernetes.io/projected/c399d3f0-061d-49b7-beb8-3bc3dc378b39-kube-api-access-9vj85\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.956084 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlv5m\" (UniqueName: \"kubernetes.io/projected/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-kube-api-access-jlv5m\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.957112 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23018ad9-0e58-4976-a215-fae5b1b51da6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.957609 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxpzv\" (UniqueName: \"kubernetes.io/projected/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-kube-api-access-kxpzv\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.957710 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-proxy-tls\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958331 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.957981 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958400 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmssh\" (UniqueName: \"kubernetes.io/projected/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-kube-api-access-mmssh\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958478 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ddea8a8-fd20-4eb2-8321-98f52f5847da-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958507 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23018ad9-0e58-4976-a215-fae5b1b51da6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958537 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ddea8a8-fd20-4eb2-8321-98f52f5847da-proxy-tls\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958516 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-trusted-ca\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958572 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgnp9\" (UniqueName: \"kubernetes.io/projected/8f91ddae-6adc-49a0-ac67-dd37606d3427-kube-api-access-dgnp9\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958601 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-metrics-certs\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.958623 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-config\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959050 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-service-ca\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959128 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-ca\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959197 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-client\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959218 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cc4744df-9673-48f7-86cb-51df34fcaf17-metrics-tls\") pod \"dns-operator-744455d44c-sbwkd\" (UID: \"cc4744df-9673-48f7-86cb-51df34fcaf17\") " pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959285 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-serving-cert\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959298 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed54eb3b-78b4-4c57-bb72-49d830aede0d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959351 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-config\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959408 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tffwc\" (UniqueName: \"kubernetes.io/projected/b0bae24b-7c92-4d48-a608-9b7a62c37161-kube-api-access-tffwc\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959430 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-images\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959460 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-srv-cert\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959485 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-profile-collector-cert\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959512 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c33e835e-eefc-4c71-84ba-a7e6784b17bd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zp4tc\" (UID: \"c33e835e-eefc-4c71-84ba-a7e6784b17bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959522 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959616 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c399d3f0-061d-49b7-beb8-3bc3dc378b39-tmpfs\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959526 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.959930 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ddea8a8-fd20-4eb2-8321-98f52f5847da-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.960030 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-images\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.960197 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c399d3f0-061d-49b7-beb8-3bc3dc378b39-tmpfs\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.961201 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-metrics-tls\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.961457 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4a6790bc-0770-4667-94d9-4cc203809743-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6tkbm\" (UID: \"4a6790bc-0770-4667-94d9-4cc203809743\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.969473 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-proxy-tls\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.971566 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.985806 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 05 06:55:09 crc kubenswrapper[4935]: I1005 06:55:09.988409 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c55e2cd9-3e55-492c-b005-d30c278bfb03-config\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.007047 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.020835 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c55e2cd9-3e55-492c-b005-d30c278bfb03-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.025998 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.033332 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cc4744df-9673-48f7-86cb-51df34fcaf17-metrics-tls\") pod \"dns-operator-744455d44c-sbwkd\" (UID: \"cc4744df-9673-48f7-86cb-51df34fcaf17\") " pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.046149 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.082305 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.089456 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.112104 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.122147 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7ddea8a8-fd20-4eb2-8321-98f52f5847da-proxy-tls\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.126147 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.145436 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.149507 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed54eb3b-78b4-4c57-bb72-49d830aede0d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.165676 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.169839 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed54eb3b-78b4-4c57-bb72-49d830aede0d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.182639 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.186775 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.206735 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.227029 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.247049 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.267965 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.278610 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0bae24b-7c92-4d48-a608-9b7a62c37161-serving-cert\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.308007 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.308323 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.317953 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-client\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.326322 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.332026 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-service-ca\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.348729 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.352311 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-etcd-ca\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.367157 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.370714 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0bae24b-7c92-4d48-a608-9b7a62c37161-config\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.386696 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.406931 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.427003 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.443121 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rphcb"] Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.448735 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.461378 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.465690 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 05 06:55:10 crc kubenswrapper[4935]: W1005 06:55:10.468291 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4064a1c6_00f1_4c14_9ce9_42308f63069f.slice/crio-e4cd27476b88e9ffd89568e18f0706564f9c667e9376dd01ec0bab355c0e93ca WatchSource:0}: Error finding container e4cd27476b88e9ffd89568e18f0706564f9c667e9376dd01ec0bab355c0e93ca: Status 404 returned error can't find the container with id e4cd27476b88e9ffd89568e18f0706564f9c667e9376dd01ec0bab355c0e93ca Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.469139 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.486126 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.507330 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.526596 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.545133 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" event={"ID":"4064a1c6-00f1-4c14-9ce9-42308f63069f","Type":"ContainerStarted","Data":"e4cd27476b88e9ffd89568e18f0706564f9c667e9376dd01ec0bab355c0e93ca"} Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.547158 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.556198 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.566862 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.570941 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-config\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.587334 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.606151 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.627294 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.637624 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23018ad9-0e58-4976-a215-fae5b1b51da6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.646233 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.667534 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.686138 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.707517 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.726599 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.744392 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-stats-auth\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.746040 4935 request.go:700] Waited for 1.00137415s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-dockercfg-zdk86&limit=500&resourceVersion=0 Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.747750 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.766568 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.774651 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-default-certificate\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.793952 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.806609 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.808922 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ef91e14-6bf7-47de-b6da-ec630e189389-service-ca-bundle\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.826212 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.846020 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.853556 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef91e14-6bf7-47de-b6da-ec630e189389-metrics-certs\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.868076 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.887277 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.898246 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-profile-collector-cert\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.899081 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-secret-volume\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.900705 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/671f9210-cd88-458e-aec5-eceaadcc3f23-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.907142 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.927195 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.940780 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/671f9210-cd88-458e-aec5-eceaadcc3f23-srv-cert\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.946121 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.956948 4935 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.956995 4935 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957008 4935 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957087 4935 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957133 4935 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957024 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8df3f651-d6cd-48a2-a97e-84448ce9fb00-package-server-manager-serving-cert podName:8df3f651-d6cd-48a2-a97e-84448ce9fb00 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.457001857 +0000 UTC m=+145.339628317 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8df3f651-d6cd-48a2-a97e-84448ce9fb00-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-r798b" (UID: "8df3f651-d6cd-48a2-a97e-84448ce9fb00") : failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957010 4935 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.956949 4935 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957185 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-cabundle podName:a0dfb3c8-3747-4a0a-b591-799bc73a7dc5 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.457160631 +0000 UTC m=+145.339787111 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-cabundle") pod "service-ca-9c57cc56f-jbxvd" (UID: "a0dfb3c8-3747-4a0a-b591-799bc73a7dc5") : failed to sync configmap cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957213 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac0a9c01-eaff-4da6-a599-48af41b70260-serving-cert podName:ac0a9c01-eaff-4da6-a599-48af41b70260 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.457200963 +0000 UTC m=+145.339827443 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/ac0a9c01-eaff-4da6-a599-48af41b70260-serving-cert") pod "service-ca-operator-777779d784-drgxn" (UID: "ac0a9c01-eaff-4da6-a599-48af41b70260") : failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957230 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-webhook-cert podName:c399d3f0-061d-49b7-beb8-3bc3dc378b39 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.457223023 +0000 UTC m=+145.339849493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-webhook-cert") pod "packageserver-d55dfcdfc-qzm8l" (UID: "c399d3f0-061d-49b7-beb8-3bc3dc378b39") : failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957249 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ac0a9c01-eaff-4da6-a599-48af41b70260-config podName:ac0a9c01-eaff-4da6-a599-48af41b70260 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.457241924 +0000 UTC m=+145.339868394 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/ac0a9c01-eaff-4da6-a599-48af41b70260-config") pod "service-ca-operator-777779d784-drgxn" (UID: "ac0a9c01-eaff-4da6-a599-48af41b70260") : failed to sync configmap cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957266 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-key podName:a0dfb3c8-3747-4a0a-b591-799bc73a7dc5 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.457258974 +0000 UTC m=+145.339885454 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-key") pod "service-ca-9c57cc56f-jbxvd" (UID: "a0dfb3c8-3747-4a0a-b591-799bc73a7dc5") : failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957285 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume podName:7ce9bae6-1db9-4a69-bef4-1f5da3ea1991 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.457274195 +0000 UTC m=+145.339900675 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume") pod "collect-profiles-29327445-sh9hj" (UID: "7ce9bae6-1db9-4a69-bef4-1f5da3ea1991") : failed to sync configmap cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957312 4935 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.957344 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-apiservice-cert podName:c399d3f0-061d-49b7-beb8-3bc3dc378b39 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.457335416 +0000 UTC m=+145.339961886 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-apiservice-cert") pod "packageserver-d55dfcdfc-qzm8l" (UID: "c399d3f0-061d-49b7-beb8-3bc3dc378b39") : failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.959195 4935 secret.go:188] Couldn't get secret openshift-kube-scheduler-operator/kube-scheduler-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.959279 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23018ad9-0e58-4976-a215-fae5b1b51da6-serving-cert podName:23018ad9-0e58-4976-a215-fae5b1b51da6 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.459256628 +0000 UTC m=+145.341883128 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/23018ad9-0e58-4976-a215-fae5b1b51da6-serving-cert") pod "openshift-kube-scheduler-operator-5fdd9b5758-sfw52" (UID: "23018ad9-0e58-4976-a215-fae5b1b51da6") : failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.960064 4935 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: E1005 06:55:10.960458 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-srv-cert podName:8f91ddae-6adc-49a0-ac67-dd37606d3427 nodeName:}" failed. No retries permitted until 2025-10-05 06:55:11.460406069 +0000 UTC m=+145.343032699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-srv-cert") pod "catalog-operator-68c6474976-fsczh" (UID: "8f91ddae-6adc-49a0-ac67-dd37606d3427") : failed to sync secret cache: timed out waiting for the condition Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.967349 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 05 06:55:10 crc kubenswrapper[4935]: I1005 06:55:10.986485 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.006380 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.026472 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.047078 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.066165 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.087598 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.105644 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.126025 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.147175 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.166926 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.186592 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.207162 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.228114 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.246370 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.267703 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.306850 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.326500 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.348589 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.376200 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.390748 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.428852 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j2nz\" (UniqueName: \"kubernetes.io/projected/c2923579-4c95-4150-9786-878a133cc0a4-kube-api-access-6j2nz\") pod \"cluster-samples-operator-665b6dd947-zglwh\" (UID: \"c2923579-4c95-4150-9786-878a133cc0a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.445964 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bgr8\" (UniqueName: \"kubernetes.io/projected/4fb78a87-5385-4785-b4b4-4e077d5000ed-kube-api-access-9bgr8\") pod \"controller-manager-879f6c89f-nh99w\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.447334 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.466128 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.490655 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkmh6\" (UniqueName: \"kubernetes.io/projected/8088d77f-c1e1-4469-af37-304d63c82f1b-kube-api-access-dkmh6\") pod \"authentication-operator-69f744f599-44xpc\" (UID: \"8088d77f-c1e1-4469-af37-304d63c82f1b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.498562 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-srv-cert\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.498731 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8df3f651-d6cd-48a2-a97e-84448ce9fb00-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r798b\" (UID: \"8df3f651-d6cd-48a2-a97e-84448ce9fb00\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.498775 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-cabundle\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.498822 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-key\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.498870 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-webhook-cert\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.498956 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac0a9c01-eaff-4da6-a599-48af41b70260-config\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.499047 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.499160 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0a9c01-eaff-4da6-a599-48af41b70260-serving-cert\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.499242 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-apiservice-cert\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.499376 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23018ad9-0e58-4976-a215-fae5b1b51da6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.501305 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-cabundle\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.501959 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac0a9c01-eaff-4da6-a599-48af41b70260-config\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.503127 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.505397 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0a9c01-eaff-4da6-a599-48af41b70260-serving-cert\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.506206 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-apiservice-cert\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.506442 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8df3f651-d6cd-48a2-a97e-84448ce9fb00-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r798b\" (UID: \"8df3f651-d6cd-48a2-a97e-84448ce9fb00\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.506459 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23018ad9-0e58-4976-a215-fae5b1b51da6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.506661 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c399d3f0-061d-49b7-beb8-3bc3dc378b39-webhook-cert\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.507169 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-signing-key\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.507564 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f91ddae-6adc-49a0-ac67-dd37606d3427-srv-cert\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.511440 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hgg5\" (UniqueName: \"kubernetes.io/projected/ac0661f5-9317-459e-bfce-c2ff9058c319-kube-api-access-4hgg5\") pod \"console-f9d7485db-cbb4c\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.513026 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.530191 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nxrl\" (UniqueName: \"kubernetes.io/projected/14973180-053d-4f7b-9f72-7acc0f8b19ac-kube-api-access-9nxrl\") pod \"oauth-openshift-558db77b4-9hwvw\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.557627 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66lb8\" (UniqueName: \"kubernetes.io/projected/7039dd6e-0052-4ea2-8a08-be11c457726d-kube-api-access-66lb8\") pod \"apiserver-7bbb656c7d-2rzp8\" (UID: \"7039dd6e-0052-4ea2-8a08-be11c457726d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.560466 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" event={"ID":"4064a1c6-00f1-4c14-9ce9-42308f63069f","Type":"ContainerStarted","Data":"6ea0b3c3133023c700e97a20db76014201a27966976a4cf054a56b71a889737a"} Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.560519 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" event={"ID":"4064a1c6-00f1-4c14-9ce9-42308f63069f","Type":"ContainerStarted","Data":"fc7359507540a6136bcd026637be0b065968a9addfb09e3e16398d051e1ec3eb"} Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.561444 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.562983 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-785ds\" (UniqueName: \"kubernetes.io/projected/512c5ef6-c7a2-4899-9db9-5e27fa61bac7-kube-api-access-785ds\") pod \"cluster-image-registry-operator-dc59b4c8b-dvk9p\" (UID: \"512c5ef6-c7a2-4899-9db9-5e27fa61bac7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.585542 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rr7r\" (UniqueName: \"kubernetes.io/projected/d90f77af-0ce4-4b2f-87d2-049e45327404-kube-api-access-8rr7r\") pod \"route-controller-manager-6576b87f9c-vwzq7\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.585766 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.606909 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.607298 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.627381 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.637045 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.647237 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.688926 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dsr4\" (UniqueName: \"kubernetes.io/projected/fb8e8feb-a6f4-4ace-a1da-a603f41feb41-kube-api-access-7dsr4\") pod \"apiserver-76f77b778f-f254r\" (UID: \"fb8e8feb-a6f4-4ace-a1da-a603f41feb41\") " pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.700047 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.700453 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.708808 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hwbx\" (UniqueName: \"kubernetes.io/projected/e7e163b1-79e1-4547-b3c9-4a506b887337-kube-api-access-5hwbx\") pod \"console-operator-58897d9998-cn4mw\" (UID: \"e7e163b1-79e1-4547-b3c9-4a506b887337\") " pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.715038 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh"] Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.721346 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.728836 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.733397 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hcpg\" (UniqueName: \"kubernetes.io/projected/3ccb7e54-6381-4dfd-92c8-679d33356e40-kube-api-access-5hcpg\") pod \"machine-approver-56656f9798-gl8x7\" (UID: \"3ccb7e54-6381-4dfd-92c8-679d33356e40\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.745760 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.748403 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6kws\" (UniqueName: \"kubernetes.io/projected/c167120c-3675-4125-a194-8a2d22959439-kube-api-access-p6kws\") pod \"openshift-apiserver-operator-796bbdcf4f-8m4vn\" (UID: \"c167120c-3675-4125-a194-8a2d22959439\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.764848 4935 request.go:700] Waited for 1.889826751s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.771222 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.771275 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.788284 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.809961 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.817800 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-44xpc"] Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.829814 4935 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.848100 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.861832 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.864816 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nh99w"] Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.867001 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 05 06:55:11 crc kubenswrapper[4935]: W1005 06:55:11.870278 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8088d77f_c1e1_4469_af37_304d63c82f1b.slice/crio-9ff8ec9274b51bcd2aa8e687eee3848b86dc1c43369a8e6cf5085111fbd1a1b8 WatchSource:0}: Error finding container 9ff8ec9274b51bcd2aa8e687eee3848b86dc1c43369a8e6cf5085111fbd1a1b8: Status 404 returned error can't find the container with id 9ff8ec9274b51bcd2aa8e687eee3848b86dc1c43369a8e6cf5085111fbd1a1b8 Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.887991 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.905794 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.920725 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.947563 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8"] Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.961925 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7"] Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.967708 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5np92\" (UniqueName: \"kubernetes.io/projected/cc4744df-9673-48f7-86cb-51df34fcaf17-kube-api-access-5np92\") pod \"dns-operator-744455d44c-sbwkd\" (UID: \"cc4744df-9673-48f7-86cb-51df34fcaf17\") " pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.984143 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d4e7a8-02e7-40b1-a024-cfe05f311c17-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-579sk\" (UID: \"c2d4e7a8-02e7-40b1-a024-cfe05f311c17\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:11 crc kubenswrapper[4935]: I1005 06:55:11.995838 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.003314 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbpg2\" (UniqueName: \"kubernetes.io/projected/671f9210-cd88-458e-aec5-eceaadcc3f23-kube-api-access-zbpg2\") pod \"olm-operator-6b444d44fb-5rxlb\" (UID: \"671f9210-cd88-458e-aec5-eceaadcc3f23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.022436 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t522c\" (UniqueName: \"kubernetes.io/projected/8ef91e14-6bf7-47de-b6da-ec630e189389-kube-api-access-t522c\") pod \"router-default-5444994796-gc2gr\" (UID: \"8ef91e14-6bf7-47de-b6da-ec630e189389\") " pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.045656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5qbp\" (UniqueName: \"kubernetes.io/projected/a0dfb3c8-3747-4a0a-b591-799bc73a7dc5-kube-api-access-j5qbp\") pod \"service-ca-9c57cc56f-jbxvd\" (UID: \"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.068055 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.068925 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvcw7\" (UniqueName: \"kubernetes.io/projected/d72e3174-d54b-4c03-90d3-8f34d7d9ce52-kube-api-access-kvcw7\") pod \"kube-storage-version-migrator-operator-b67b599dd-7zklk\" (UID: \"d72e3174-d54b-4c03-90d3-8f34d7d9ce52\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.076684 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9hwvw"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.084956 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qh47\" (UniqueName: \"kubernetes.io/projected/8df3f651-d6cd-48a2-a97e-84448ce9fb00-kube-api-access-8qh47\") pod \"package-server-manager-789f6589d5-r798b\" (UID: \"8df3f651-d6cd-48a2-a97e-84448ce9fb00\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:12 crc kubenswrapper[4935]: W1005 06:55:12.097289 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14973180_053d_4f7b_9f72_7acc0f8b19ac.slice/crio-559dae40682f56604252d745d43ba83d9e5d57e959267f82229aac17941f4d53 WatchSource:0}: Error finding container 559dae40682f56604252d745d43ba83d9e5d57e959267f82229aac17941f4d53: Status 404 returned error can't find the container with id 559dae40682f56604252d745d43ba83d9e5d57e959267f82229aac17941f4d53 Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.102701 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwwgj\" (UniqueName: \"kubernetes.io/projected/7ddea8a8-fd20-4eb2-8321-98f52f5847da-kube-api-access-gwwgj\") pod \"machine-config-controller-84d6567774-hkqkc\" (UID: \"7ddea8a8-fd20-4eb2-8321-98f52f5847da\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.108529 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.117703 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.124314 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bvx8\" (UniqueName: \"kubernetes.io/projected/c1af3cc3-3fad-473d-a193-7de470b5ff7c-kube-api-access-6bvx8\") pod \"migrator-59844c95c7-7ckxt\" (UID: \"c1af3cc3-3fad-473d-a193-7de470b5ff7c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.140224 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.141354 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-bound-sa-token\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.144960 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.177471 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vj85\" (UniqueName: \"kubernetes.io/projected/c399d3f0-061d-49b7-beb8-3bc3dc378b39-kube-api-access-9vj85\") pod \"packageserver-d55dfcdfc-qzm8l\" (UID: \"c399d3f0-061d-49b7-beb8-3bc3dc378b39\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.181555 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c26rq\" (UniqueName: \"kubernetes.io/projected/ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70-kube-api-access-c26rq\") pod \"ingress-operator-5b745b69d9-trqss\" (UID: \"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.187063 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.204874 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cn4mw"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.209324 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhvn2\" (UniqueName: \"kubernetes.io/projected/4a6790bc-0770-4667-94d9-4cc203809743-kube-api-access-nhvn2\") pod \"multus-admission-controller-857f4d67dd-6tkbm\" (UID: \"4a6790bc-0770-4667-94d9-4cc203809743\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.215748 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.224831 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlv5m\" (UniqueName: \"kubernetes.io/projected/42b7b98d-c2b9-4a61-aabc-e6ec081a3d06-kube-api-access-jlv5m\") pod \"machine-config-operator-74547568cd-wcrn7\" (UID: \"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:12 crc kubenswrapper[4935]: W1005 06:55:12.232107 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7e163b1_79e1_4547_b3c9_4a506b887337.slice/crio-fe6c07f1d4e2b5b09d0217bfccf8cf475d6b6d01339f1a27acc058b9791f8707 WatchSource:0}: Error finding container fe6c07f1d4e2b5b09d0217bfccf8cf475d6b6d01339f1a27acc058b9791f8707: Status 404 returned error can't find the container with id fe6c07f1d4e2b5b09d0217bfccf8cf475d6b6d01339f1a27acc058b9791f8707 Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.243959 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnjdm\" (UniqueName: \"kubernetes.io/projected/ed54eb3b-78b4-4c57-bb72-49d830aede0d-kube-api-access-hnjdm\") pod \"openshift-controller-manager-operator-756b6f6bc6-n2pnb\" (UID: \"ed54eb3b-78b4-4c57-bb72-49d830aede0d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:12 crc kubenswrapper[4935]: W1005 06:55:12.259437 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ef91e14_6bf7_47de_b6da_ec630e189389.slice/crio-372c831623388742a3e967b0916349c22fcf94869fdf1f488cc9da30ea31a446 WatchSource:0}: Error finding container 372c831623388742a3e967b0916349c22fcf94869fdf1f488cc9da30ea31a446: Status 404 returned error can't find the container with id 372c831623388742a3e967b0916349c22fcf94869fdf1f488cc9da30ea31a446 Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.263231 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4shb7\" (UniqueName: \"kubernetes.io/projected/ac0a9c01-eaff-4da6-a599-48af41b70260-kube-api-access-4shb7\") pod \"service-ca-operator-777779d784-drgxn\" (UID: \"ac0a9c01-eaff-4da6-a599-48af41b70260\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.291451 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23018ad9-0e58-4976-a215-fae5b1b51da6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sfw52\" (UID: \"23018ad9-0e58-4976-a215-fae5b1b51da6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.305277 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxpzv\" (UniqueName: \"kubernetes.io/projected/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-kube-api-access-kxpzv\") pod \"collect-profiles-29327445-sh9hj\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.312113 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f254r"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.316579 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.322852 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c55e2cd9-3e55-492c-b005-d30c278bfb03-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-468br\" (UID: \"c55e2cd9-3e55-492c-b005-d30c278bfb03\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.328993 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.329777 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-cbb4c"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.335584 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.343391 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zfm6\" (UniqueName: \"kubernetes.io/projected/c33e835e-eefc-4c71-84ba-a7e6784b17bd-kube-api-access-9zfm6\") pod \"control-plane-machine-set-operator-78cbb6b69f-zp4tc\" (UID: \"c33e835e-eefc-4c71-84ba-a7e6784b17bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.343943 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.349069 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.363026 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.376279 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsn2q\" (UniqueName: \"kubernetes.io/projected/22367e16-ef10-498d-8cad-b719508ea6eb-kube-api-access-lsn2q\") pod \"downloads-7954f5f757-597nr\" (UID: \"22367e16-ef10-498d-8cad-b719508ea6eb\") " pod="openshift-console/downloads-7954f5f757-597nr" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.377218 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.377525 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-597nr" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.384808 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.386481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmssh\" (UniqueName: \"kubernetes.io/projected/9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7-kube-api-access-mmssh\") pod \"openshift-config-operator-7777fb866f-gx7h8\" (UID: \"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.390832 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.409988 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgnp9\" (UniqueName: \"kubernetes.io/projected/8f91ddae-6adc-49a0-ac67-dd37606d3427-kube-api-access-dgnp9\") pod \"catalog-operator-68c6474976-fsczh\" (UID: \"8f91ddae-6adc-49a0-ac67-dd37606d3427\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.423046 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tffwc\" (UniqueName: \"kubernetes.io/projected/b0bae24b-7c92-4d48-a608-9b7a62c37161-kube-api-access-tffwc\") pod \"etcd-operator-b45778765-mzckv\" (UID: \"b0bae24b-7c92-4d48-a608-9b7a62c37161\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.423745 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.452265 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.465425 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.500202 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.519479 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.524770 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.525453 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-certificates\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.525482 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a34ebce5-75e3-4033-9fa0-bb17b2151fae-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.525516 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a34ebce5-75e3-4033-9fa0-bb17b2151fae-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.525532 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-bound-sa-token\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.525549 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-tls\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.525683 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gxhl\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-kube-api-access-5gxhl\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.525704 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-trusted-ca\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.525727 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: E1005 06:55:12.525996 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.025983226 +0000 UTC m=+146.908609686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.596659 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jbxvd"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.603589 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cn4mw" event={"ID":"e7e163b1-79e1-4547-b3c9-4a506b887337","Type":"ContainerStarted","Data":"3ef5e725d54569f64fcd44a1bc761f8d0400ec3d9576241283cb5f3a07959e00"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.603980 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cn4mw" event={"ID":"e7e163b1-79e1-4547-b3c9-4a506b887337","Type":"ContainerStarted","Data":"fe6c07f1d4e2b5b09d0217bfccf8cf475d6b6d01339f1a27acc058b9791f8707"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.607341 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.608261 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.614365 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.615049 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" event={"ID":"3ccb7e54-6381-4dfd-92c8-679d33356e40","Type":"ContainerStarted","Data":"75996b8e6b2831a30ad373694be681839c1dcfc60b244e06587f90361644c30f"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.615088 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" event={"ID":"3ccb7e54-6381-4dfd-92c8-679d33356e40","Type":"ContainerStarted","Data":"81503f4dddc41969301b515a0992062d6687257c768b9fffa78c88354c2c96af"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.616282 4935 patch_prober.go:28] interesting pod/console-operator-58897d9998-cn4mw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.616348 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-cn4mw" podUID="e7e163b1-79e1-4547-b3c9-4a506b887337" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.619512 4935 generic.go:334] "Generic (PLEG): container finished" podID="7039dd6e-0052-4ea2-8a08-be11c457726d" containerID="ffea892229ef0c37bb7381a765166827ef75dff75ae1961de407ab791c5c2263" exitCode=0 Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.619564 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" event={"ID":"7039dd6e-0052-4ea2-8a08-be11c457726d","Type":"ContainerDied","Data":"ffea892229ef0c37bb7381a765166827ef75dff75ae1961de407ab791c5c2263"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.619583 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" event={"ID":"7039dd6e-0052-4ea2-8a08-be11c457726d","Type":"ContainerStarted","Data":"259b5a8d642fc77125a1ee48c675eaf79e2b9bad55e3ffed011d12ee39355c2e"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.624823 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cbb4c" event={"ID":"ac0661f5-9317-459e-bfce-c2ff9058c319","Type":"ContainerStarted","Data":"0117757e0089912a92e1f34ca9432a4c2463541b9dd29cb07f60d5c862c29738"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.626713 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:12 crc kubenswrapper[4935]: E1005 06:55:12.626812 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.126786517 +0000 UTC m=+147.009412977 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.626857 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-trusted-ca\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.626912 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.626934 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84bst\" (UniqueName: \"kubernetes.io/projected/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-kube-api-access-84bst\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627032 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-certificates\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627054 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-csi-data-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627104 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-certs\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627205 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw47t\" (UniqueName: \"kubernetes.io/projected/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-kube-api-access-mw47t\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627256 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a34ebce5-75e3-4033-9fa0-bb17b2151fae-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627295 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab0901a2-e6ff-44a3-8382-d85f25eeba64-cert\") pod \"ingress-canary-68swt\" (UID: \"ab0901a2-e6ff-44a3-8382-d85f25eeba64\") " pod="openshift-ingress-canary/ingress-canary-68swt" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627348 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5rsz\" (UniqueName: \"kubernetes.io/projected/ab0901a2-e6ff-44a3-8382-d85f25eeba64-kube-api-access-c5rsz\") pod \"ingress-canary-68swt\" (UID: \"ab0901a2-e6ff-44a3-8382-d85f25eeba64\") " pod="openshift-ingress-canary/ingress-canary-68swt" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627394 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-registration-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627446 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-socket-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627504 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-metrics-tls\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627520 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-mountpoint-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627557 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a34ebce5-75e3-4033-9fa0-bb17b2151fae-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627574 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-bound-sa-token\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.627622 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-tls\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.628023 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.628051 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s9dn\" (UniqueName: \"kubernetes.io/projected/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-kube-api-access-9s9dn\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.628081 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-config-volume\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.628104 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gt6n\" (UniqueName: \"kubernetes.io/projected/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-kube-api-access-7gt6n\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.628282 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-plugins-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.628366 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gxhl\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-kube-api-access-5gxhl\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.628388 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.628407 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-node-bootstrap-token\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.633599 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a34ebce5-75e3-4033-9fa0-bb17b2151fae-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.636232 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" Oct 05 06:55:12 crc kubenswrapper[4935]: E1005 06:55:12.637821 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.137804776 +0000 UTC m=+147.020431236 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.640320 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-trusted-ca\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.641489 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sbwkd"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.653730 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a34ebce5-75e3-4033-9fa0-bb17b2151fae-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.654061 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-certificates\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.656097 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-tls\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.659697 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" event={"ID":"d90f77af-0ce4-4b2f-87d2-049e45327404","Type":"ContainerStarted","Data":"068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.659734 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" event={"ID":"d90f77af-0ce4-4b2f-87d2-049e45327404","Type":"ContainerStarted","Data":"69c3c2cc36de60ce4ea8294fa1bf47b0397e81a0d64a1cd7cb3529d39d08279b"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.660024 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.660428 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.662612 4935 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-vwzq7 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.662651 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" podUID="d90f77af-0ce4-4b2f-87d2-049e45327404" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.670253 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gc2gr" event={"ID":"8ef91e14-6bf7-47de-b6da-ec630e189389","Type":"ContainerStarted","Data":"b6a0c9c327eb8949715768f7dd3b190a9a6bd0303b1a2de9395e6770a730f7a6"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.670295 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gc2gr" event={"ID":"8ef91e14-6bf7-47de-b6da-ec630e189389","Type":"ContainerStarted","Data":"372c831623388742a3e967b0916349c22fcf94869fdf1f488cc9da30ea31a446"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.676884 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f254r" event={"ID":"fb8e8feb-a6f4-4ace-a1da-a603f41feb41","Type":"ContainerStarted","Data":"2de7103a12f6057513aa3c8f2919a40c679d93db7678e25f4ee6c5396bd3d461"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.680794 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" event={"ID":"c167120c-3675-4125-a194-8a2d22959439","Type":"ContainerStarted","Data":"d08f1f936757c6700ab327e2418b35e3dc33119269a774f8c5e6192550fbd9fc"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.685055 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" event={"ID":"c2923579-4c95-4150-9786-878a133cc0a4","Type":"ContainerStarted","Data":"2bd01a9f984ef12d9ea6782c2db66d79451f4320714423443cc12376771c4369"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.685095 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" event={"ID":"c2923579-4c95-4150-9786-878a133cc0a4","Type":"ContainerStarted","Data":"f16e6ed45f9f71dc194393d2cc442e4ea4882270d37dc4d1b0648b5b6525a7bd"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.685110 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" event={"ID":"c2923579-4c95-4150-9786-878a133cc0a4","Type":"ContainerStarted","Data":"ecc06200942fc462a2cb7e2a16cc0f065a74f04f2088ff4aeb8f748eb91b0871"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.685728 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-bound-sa-token\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.694036 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" event={"ID":"512c5ef6-c7a2-4899-9db9-5e27fa61bac7","Type":"ContainerStarted","Data":"35e3434879f81e78b4927fd98b87b66c528472afd1fdc2c66698a84572b1352b"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.694088 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" event={"ID":"512c5ef6-c7a2-4899-9db9-5e27fa61bac7","Type":"ContainerStarted","Data":"8f59432a78beb8f492f924232e36c35cbeaae198ccc306f77959dada46cedebd"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.700950 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.702701 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.706225 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gxhl\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-kube-api-access-5gxhl\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: W1005 06:55:12.717756 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0dfb3c8_3747_4a0a_b591_799bc73a7dc5.slice/crio-efd130d781abdf75123d9f8ef5fa80c753f2b511f728dae4e2ce5e7505c002f5 WatchSource:0}: Error finding container efd130d781abdf75123d9f8ef5fa80c753f2b511f728dae4e2ce5e7505c002f5: Status 404 returned error can't find the container with id efd130d781abdf75123d9f8ef5fa80c753f2b511f728dae4e2ce5e7505c002f5 Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.719503 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" event={"ID":"4fb78a87-5385-4785-b4b4-4e077d5000ed","Type":"ContainerStarted","Data":"2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.719546 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" event={"ID":"4fb78a87-5385-4785-b4b4-4e077d5000ed","Type":"ContainerStarted","Data":"d3de02116d55c7f1b4eda70beff2e4443e2e247134aa171eb917401adbf0390e"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.720327 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.723034 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" event={"ID":"d72e3174-d54b-4c03-90d3-8f34d7d9ce52","Type":"ContainerStarted","Data":"7fdde0626cc636d6fa361dc8d6a1aef6d87fb3a6e1ef2144260f29176c72c289"} Oct 05 06:55:12 crc kubenswrapper[4935]: W1005 06:55:12.723999 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2d4e7a8_02e7_40b1_a024_cfe05f311c17.slice/crio-762d363feada9aec9dc6e4c432775c59fd2487dbe67e136190a1e8e83fe2b10d WatchSource:0}: Error finding container 762d363feada9aec9dc6e4c432775c59fd2487dbe67e136190a1e8e83fe2b10d: Status 404 returned error can't find the container with id 762d363feada9aec9dc6e4c432775c59fd2487dbe67e136190a1e8e83fe2b10d Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.731756 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732480 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84bst\" (UniqueName: \"kubernetes.io/projected/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-kube-api-access-84bst\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732541 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-csi-data-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732574 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-certs\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732608 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw47t\" (UniqueName: \"kubernetes.io/projected/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-kube-api-access-mw47t\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732653 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab0901a2-e6ff-44a3-8382-d85f25eeba64-cert\") pod \"ingress-canary-68swt\" (UID: \"ab0901a2-e6ff-44a3-8382-d85f25eeba64\") " pod="openshift-ingress-canary/ingress-canary-68swt" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732688 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5rsz\" (UniqueName: \"kubernetes.io/projected/ab0901a2-e6ff-44a3-8382-d85f25eeba64-kube-api-access-c5rsz\") pod \"ingress-canary-68swt\" (UID: \"ab0901a2-e6ff-44a3-8382-d85f25eeba64\") " pod="openshift-ingress-canary/ingress-canary-68swt" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732715 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-registration-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732758 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-socket-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732778 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-metrics-tls\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732799 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-mountpoint-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732932 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.732988 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s9dn\" (UniqueName: \"kubernetes.io/projected/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-kube-api-access-9s9dn\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.733012 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-config-volume\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.733032 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gt6n\" (UniqueName: \"kubernetes.io/projected/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-kube-api-access-7gt6n\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.733572 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-plugins-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.733595 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.733618 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-node-bootstrap-token\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.741049 4935 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nh99w container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.741134 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" podUID="4fb78a87-5385-4785-b4b4-4e077d5000ed" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 05 06:55:12 crc kubenswrapper[4935]: E1005 06:55:12.742462 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.242441311 +0000 UTC m=+147.125067771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.742979 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-csi-data-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.743176 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" event={"ID":"8088d77f-c1e1-4469-af37-304d63c82f1b","Type":"ContainerStarted","Data":"3c5f1d6ff008ddd2ed097fbaa5f6c68d9fd6131169ff901fda112fefce68496c"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.743247 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" event={"ID":"8088d77f-c1e1-4469-af37-304d63c82f1b","Type":"ContainerStarted","Data":"9ff8ec9274b51bcd2aa8e687eee3848b86dc1c43369a8e6cf5085111fbd1a1b8"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.746810 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.747095 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-socket-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.747162 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-config-volume\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.748022 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-mountpoint-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.748599 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-plugins-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.748922 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-registration-dir\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.751031 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" event={"ID":"14973180-053d-4f7b-9f72-7acc0f8b19ac","Type":"ContainerStarted","Data":"7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.751071 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" event={"ID":"14973180-053d-4f7b-9f72-7acc0f8b19ac","Type":"ContainerStarted","Data":"559dae40682f56604252d745d43ba83d9e5d57e959267f82229aac17941f4d53"} Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.751936 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.753726 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-certs\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.753796 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ab0901a2-e6ff-44a3-8382-d85f25eeba64-cert\") pod \"ingress-canary-68swt\" (UID: \"ab0901a2-e6ff-44a3-8382-d85f25eeba64\") " pod="openshift-ingress-canary/ingress-canary-68swt" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.756536 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.756734 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-metrics-tls\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.759775 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-node-bootstrap-token\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.763987 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6tkbm"] Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.765995 4935 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9hwvw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.766052 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" podUID="14973180-053d-4f7b-9f72-7acc0f8b19ac" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.787745 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84bst\" (UniqueName: \"kubernetes.io/projected/6a92699e-19a9-409a-b5e8-b79fc23b5d1b-kube-api-access-84bst\") pod \"machine-config-server-zdg8q\" (UID: \"6a92699e-19a9-409a-b5e8-b79fc23b5d1b\") " pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.805523 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gt6n\" (UniqueName: \"kubernetes.io/projected/e0336e4e-ef2a-42f5-838c-c8e6dd919d59-kube-api-access-7gt6n\") pod \"csi-hostpathplugin-ssjkm\" (UID: \"e0336e4e-ef2a-42f5-838c-c8e6dd919d59\") " pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.835336 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:12 crc kubenswrapper[4935]: E1005 06:55:12.837422 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.337403264 +0000 UTC m=+147.220029724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.842004 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw47t\" (UniqueName: \"kubernetes.io/projected/2b54e92f-18aa-4725-a1ec-3b3ef26e6972-kube-api-access-mw47t\") pod \"dns-default-bskzq\" (UID: \"2b54e92f-18aa-4725-a1ec-3b3ef26e6972\") " pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.842152 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zdg8q" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.850911 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s9dn\" (UniqueName: \"kubernetes.io/projected/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-kube-api-access-9s9dn\") pod \"marketplace-operator-79b997595-cvn45\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.863332 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.869862 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.869958 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5rsz\" (UniqueName: \"kubernetes.io/projected/ab0901a2-e6ff-44a3-8382-d85f25eeba64-kube-api-access-c5rsz\") pod \"ingress-canary-68swt\" (UID: \"ab0901a2-e6ff-44a3-8382-d85f25eeba64\") " pod="openshift-ingress-canary/ingress-canary-68swt" Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.936592 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:12 crc kubenswrapper[4935]: E1005 06:55:12.939444 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.438839422 +0000 UTC m=+147.321465882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:12 crc kubenswrapper[4935]: I1005 06:55:12.940037 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-cn4mw" podStartSLOduration=126.940016664 podStartE2EDuration="2m6.940016664s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:12.9395029 +0000 UTC m=+146.822129360" watchObservedRunningTime="2025-10-05 06:55:12.940016664 +0000 UTC m=+146.822643124" Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.041989 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.043126 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.543112307 +0000 UTC m=+147.425738767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.128709 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.138860 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-68swt" Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.148177 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.148698 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.648676647 +0000 UTC m=+147.531303107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.149105 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.149608 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.149640 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.196223 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.251367 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.252009 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.751962604 +0000 UTC m=+147.634589064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: W1005 06:55:13.284975 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a92699e_19a9_409a_b5e8_b79fc23b5d1b.slice/crio-4f6e0d45c181a4b879638dc153f472d6fb7fb17d6532f711e41ca45bebeec4c6 WatchSource:0}: Error finding container 4f6e0d45c181a4b879638dc153f472d6fb7fb17d6532f711e41ca45bebeec4c6: Status 404 returned error can't find the container with id 4f6e0d45c181a4b879638dc153f472d6fb7fb17d6532f711e41ca45bebeec4c6 Oct 05 06:55:13 crc kubenswrapper[4935]: W1005 06:55:13.297855 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42b7b98d_c2b9_4a61_aabc_e6ec081a3d06.slice/crio-9d901daa16e76be72c6e91c72901534f531856d5085e06de67f2e0cdc25e1f43 WatchSource:0}: Error finding container 9d901daa16e76be72c6e91c72901534f531856d5085e06de67f2e0cdc25e1f43: Status 404 returned error can't find the container with id 9d901daa16e76be72c6e91c72901534f531856d5085e06de67f2e0cdc25e1f43 Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.344595 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-rphcb" podStartSLOduration=127.344579354 podStartE2EDuration="2m7.344579354s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:13.340579185 +0000 UTC m=+147.223205645" watchObservedRunningTime="2025-10-05 06:55:13.344579354 +0000 UTC m=+147.227205814" Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.358048 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.358920 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.858899652 +0000 UTC m=+147.741526112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.359066 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.359380 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.859372814 +0000 UTC m=+147.741999274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.389467 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.393140 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.394613 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-trqss"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.432783 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.460207 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.460438 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.960408662 +0000 UTC m=+147.843035122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.460567 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.460909 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:13.960900895 +0000 UTC m=+147.843527355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.543421 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-597nr"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.550928 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.572920 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.573213 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.073197008 +0000 UTC m=+147.955823468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.594672 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-drgxn"] Oct 05 06:55:13 crc kubenswrapper[4935]: W1005 06:55:13.624005 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23018ad9_0e58_4976_a215_fae5b1b51da6.slice/crio-8d36e9f1221a60be1373c7d45289c24033994d65c1a65f644fc24b5222d6762a WatchSource:0}: Error finding container 8d36e9f1221a60be1373c7d45289c24033994d65c1a65f644fc24b5222d6762a: Status 404 returned error can't find the container with id 8d36e9f1221a60be1373c7d45289c24033994d65c1a65f644fc24b5222d6762a Oct 05 06:55:13 crc kubenswrapper[4935]: W1005 06:55:13.635858 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac0a9c01_eaff_4da6_a599_48af41b70260.slice/crio-eef92d8d1893e9473e86f43d35382e453e9eb8fb0f6f79e900fa87d38370a23d WatchSource:0}: Error finding container eef92d8d1893e9473e86f43d35382e453e9eb8fb0f6f79e900fa87d38370a23d: Status 404 returned error can't find the container with id eef92d8d1893e9473e86f43d35382e453e9eb8fb0f6f79e900fa87d38370a23d Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.668781 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mzckv"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.678550 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.678839 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.17882605 +0000 UTC m=+148.061452510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.732212 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.735327 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc"] Oct 05 06:55:13 crc kubenswrapper[4935]: W1005 06:55:13.774337 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ce9bae6_1db9_4a69_bef4_1f5da3ea1991.slice/crio-a20cb86af93a65ffc01450017aee4ac388c8028622f5ee142d5defd547d35957 WatchSource:0}: Error finding container a20cb86af93a65ffc01450017aee4ac388c8028622f5ee142d5defd547d35957: Status 404 returned error can't find the container with id a20cb86af93a65ffc01450017aee4ac388c8028622f5ee142d5defd547d35957 Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.780813 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.781077 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.281060989 +0000 UTC m=+148.163687449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.795766 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" event={"ID":"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5","Type":"ContainerStarted","Data":"85cc59faa7e03e40e403118b887355a0239f67a82277afa99786552490aa236d"} Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.795805 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" event={"ID":"a0dfb3c8-3747-4a0a-b591-799bc73a7dc5","Type":"ContainerStarted","Data":"efd130d781abdf75123d9f8ef5fa80c753f2b511f728dae4e2ce5e7505c002f5"} Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.807844 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.830569 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.830989 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.870687 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bskzq"] Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.871873 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" event={"ID":"ac0a9c01-eaff-4da6-a599-48af41b70260","Type":"ContainerStarted","Data":"eef92d8d1893e9473e86f43d35382e453e9eb8fb0f6f79e900fa87d38370a23d"} Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.899167 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:13 crc kubenswrapper[4935]: E1005 06:55:13.900441 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.400405283 +0000 UTC m=+148.283031743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:13 crc kubenswrapper[4935]: I1005 06:55:13.947711 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ssjkm"] Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.006583 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" event={"ID":"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70","Type":"ContainerStarted","Data":"f4f3e23a9661878d8f25dc3b3faf912e9fca421621e581bf38a642e4f66c9fda"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.006962 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.009020 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.509001685 +0000 UTC m=+148.391628145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.034982 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cbb4c" event={"ID":"ac0661f5-9317-459e-bfce-c2ff9058c319","Type":"ContainerStarted","Data":"c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e"} Oct 05 06:55:14 crc kubenswrapper[4935]: W1005 06:55:14.036581 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fa12988_6e6d_4a16_9cca_3b96a4e4fbb7.slice/crio-d7f20fe688668ac0ffbef28f6104feb0cf02e95e3386ee0240a8534916e8a844 WatchSource:0}: Error finding container d7f20fe688668ac0ffbef28f6104feb0cf02e95e3386ee0240a8534916e8a844: Status 404 returned error can't find the container with id d7f20fe688668ac0ffbef28f6104feb0cf02e95e3386ee0240a8534916e8a844 Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.061211 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" event={"ID":"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06","Type":"ContainerStarted","Data":"9d901daa16e76be72c6e91c72901534f531856d5085e06de67f2e0cdc25e1f43"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.063426 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvn45"] Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.064567 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" event={"ID":"cc4744df-9673-48f7-86cb-51df34fcaf17","Type":"ContainerStarted","Data":"f151b7e83abf4b68ee25b798b8e688e73ab5f490ab328f008d07996193e9cb48"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.066262 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" event={"ID":"8df3f651-d6cd-48a2-a97e-84448ce9fb00","Type":"ContainerStarted","Data":"671ff36d389efd5f063a081c49701557259fffc26e099110dcbcdd1c0bb7ea0f"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.066296 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" event={"ID":"8df3f651-d6cd-48a2-a97e-84448ce9fb00","Type":"ContainerStarted","Data":"6b0bc0a4f25ea5fe24cfdaec1b333a3418dff5c297bea5c581e3a8de1e18c2dd"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.078492 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-68swt"] Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.101090 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zdg8q" event={"ID":"6a92699e-19a9-409a-b5e8-b79fc23b5d1b","Type":"ContainerStarted","Data":"4f6e0d45c181a4b879638dc153f472d6fb7fb17d6532f711e41ca45bebeec4c6"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.113089 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.115108 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.615088739 +0000 UTC m=+148.497715199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.123629 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" event={"ID":"3ccb7e54-6381-4dfd-92c8-679d33356e40","Type":"ContainerStarted","Data":"293fcc20c8c961400f88512f837985a25c57e0818442e02882a0c3b6d6b3c52f"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.135513 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" event={"ID":"c2d4e7a8-02e7-40b1-a024-cfe05f311c17","Type":"ContainerStarted","Data":"762d363feada9aec9dc6e4c432775c59fd2487dbe67e136190a1e8e83fe2b10d"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.148582 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:14 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:14 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:14 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.148657 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.149350 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" event={"ID":"671f9210-cd88-458e-aec5-eceaadcc3f23","Type":"ContainerStarted","Data":"042c9649145c15e80debebea8ef7c46ed19926c0f361a0d6ba5f20fa69db35d2"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.149413 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" event={"ID":"671f9210-cd88-458e-aec5-eceaadcc3f23","Type":"ContainerStarted","Data":"e177e9a431fb83fa36196349285817681dc06e2373881d9a7fff2239a7f11642"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.152084 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.152265 4935 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5rxlb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.152289 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" podUID="671f9210-cd88-458e-aec5-eceaadcc3f23" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.169240 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-597nr" event={"ID":"22367e16-ef10-498d-8cad-b719508ea6eb","Type":"ContainerStarted","Data":"9cefe31596622641e9d9dd618bbe2bdaefb76844071c186e988c9a19538a2a19"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.200430 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" event={"ID":"c167120c-3675-4125-a194-8a2d22959439","Type":"ContainerStarted","Data":"c408e9051c93f99897013bd19c22399fd8f24a4d49392f67a2b06027d06230db"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.213720 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.214242 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.714228255 +0000 UTC m=+148.596854715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.214405 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.215163 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.71512496 +0000 UTC m=+148.597751410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.223121 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" event={"ID":"4a6790bc-0770-4667-94d9-4cc203809743","Type":"ContainerStarted","Data":"dc26c85214ebcad9997bf3861e35c0ef8f84118775a837f3fb300470fb438127"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.235189 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" event={"ID":"23018ad9-0e58-4976-a215-fae5b1b51da6","Type":"ContainerStarted","Data":"8d36e9f1221a60be1373c7d45289c24033994d65c1a65f644fc24b5222d6762a"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.243855 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" event={"ID":"c55e2cd9-3e55-492c-b005-d30c278bfb03","Type":"ContainerStarted","Data":"696ce070ed1a2502663eda411dd8e601c64911b62a612c56e9088dd5ce05c620"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.251147 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" event={"ID":"ed54eb3b-78b4-4c57-bb72-49d830aede0d","Type":"ContainerStarted","Data":"cce84ee3d2518fc42c93ff95f90feaab1e24086cbd9dc70ceb30da653168568e"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.254195 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" event={"ID":"c1af3cc3-3fad-473d-a193-7de470b5ff7c","Type":"ContainerStarted","Data":"19a87fb956233e32c6d7f4b89a63f4f7386de21e6777ba4c46b997845c1d9181"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.254226 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" event={"ID":"c1af3cc3-3fad-473d-a193-7de470b5ff7c","Type":"ContainerStarted","Data":"70ffdc8baba26b6f465ce94423bfc8bbc7d4f4f65937cc4b21534ee99b051c3a"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.258827 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" event={"ID":"d72e3174-d54b-4c03-90d3-8f34d7d9ce52","Type":"ContainerStarted","Data":"8d348403248775640b56c7ad79eb9a96a62724914cbaeead1b55fe0e4bcc58c7"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.261234 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-gc2gr" podStartSLOduration=128.261218499 podStartE2EDuration="2m8.261218499s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.259118502 +0000 UTC m=+148.141744962" watchObservedRunningTime="2025-10-05 06:55:14.261218499 +0000 UTC m=+148.143844949" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.269684 4935 generic.go:334] "Generic (PLEG): container finished" podID="fb8e8feb-a6f4-4ace-a1da-a603f41feb41" containerID="518159cccfb677ca69586af370c6fe09a9dc6b13c027a327f4fb81c4916a0403" exitCode=0 Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.270176 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f254r" event={"ID":"fb8e8feb-a6f4-4ace-a1da-a603f41feb41","Type":"ContainerDied","Data":"518159cccfb677ca69586af370c6fe09a9dc6b13c027a327f4fb81c4916a0403"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.289479 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.289632 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.306926 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" event={"ID":"7ddea8a8-fd20-4eb2-8321-98f52f5847da","Type":"ContainerStarted","Data":"785ca8384a9356d1907c1f479b2801e990fdf3efab3f9b851eb2cb2ac53a8fc8"} Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.315850 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.318571 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.321547 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.323575 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.823558488 +0000 UTC m=+148.706184948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.331836 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.333572 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.833540588 +0000 UTC m=+148.716167048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.338379 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.346418 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" podStartSLOduration=128.346396276 podStartE2EDuration="2m8.346396276s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.344793313 +0000 UTC m=+148.227419773" watchObservedRunningTime="2025-10-05 06:55:14.346396276 +0000 UTC m=+148.229022736" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.369637 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" podStartSLOduration=128.369616585 podStartE2EDuration="2m8.369616585s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.30446014 +0000 UTC m=+148.187086600" watchObservedRunningTime="2025-10-05 06:55:14.369616585 +0000 UTC m=+148.252243045" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.405759 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gl8x7" podStartSLOduration=128.405735334 podStartE2EDuration="2m8.405735334s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.38822442 +0000 UTC m=+148.270850880" watchObservedRunningTime="2025-10-05 06:55:14.405735334 +0000 UTC m=+148.288361794" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.433789 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.434138 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.934092732 +0000 UTC m=+148.816719192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.434440 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.436265 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:14.936254341 +0000 UTC m=+148.818880801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.474267 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-44xpc" podStartSLOduration=128.47425081 podStartE2EDuration="2m8.47425081s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.472638627 +0000 UTC m=+148.355265097" watchObservedRunningTime="2025-10-05 06:55:14.47425081 +0000 UTC m=+148.356877270" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.535535 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.536110 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.036093976 +0000 UTC m=+148.918720436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.593203 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zglwh" podStartSLOduration=128.593185503 podStartE2EDuration="2m8.593185503s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.556343825 +0000 UTC m=+148.438970285" watchObservedRunningTime="2025-10-05 06:55:14.593185503 +0000 UTC m=+148.475811963" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.632628 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" podStartSLOduration=128.632613501 podStartE2EDuration="2m8.632613501s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.593331287 +0000 UTC m=+148.475957747" watchObservedRunningTime="2025-10-05 06:55:14.632613501 +0000 UTC m=+148.515239961" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.633148 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-jbxvd" podStartSLOduration=128.633144525 podStartE2EDuration="2m8.633144525s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.631005417 +0000 UTC m=+148.513631877" watchObservedRunningTime="2025-10-05 06:55:14.633144525 +0000 UTC m=+148.515770985" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.636668 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.636769 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.636863 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.637183 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.137168664 +0000 UTC m=+149.019795124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.638534 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.652137 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.727525 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8m4vn" podStartSLOduration=128.727508682 podStartE2EDuration="2m8.727508682s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.722748983 +0000 UTC m=+148.605375453" watchObservedRunningTime="2025-10-05 06:55:14.727508682 +0000 UTC m=+148.610135142" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.730318 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dvk9p" podStartSLOduration=128.730306878 podStartE2EDuration="2m8.730306878s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.697336404 +0000 UTC m=+148.579962874" watchObservedRunningTime="2025-10-05 06:55:14.730306878 +0000 UTC m=+148.612933328" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.737471 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.737822 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.737857 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.742004 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.241967884 +0000 UTC m=+149.124594344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.759058 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.770342 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.801883 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.819285 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.831680 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.838578 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-cbb4c" podStartSLOduration=128.838564081 podStartE2EDuration="2m8.838564081s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.791619599 +0000 UTC m=+148.674246079" watchObservedRunningTime="2025-10-05 06:55:14.838564081 +0000 UTC m=+148.721190541" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.840121 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" podStartSLOduration=128.840115413 podStartE2EDuration="2m8.840115413s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.837812391 +0000 UTC m=+148.720438851" watchObservedRunningTime="2025-10-05 06:55:14.840115413 +0000 UTC m=+148.722741873" Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.840692 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.842302 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.342284692 +0000 UTC m=+149.224911152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:14 crc kubenswrapper[4935]: I1005 06:55:14.952348 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:14 crc kubenswrapper[4935]: E1005 06:55:14.952690 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.452662782 +0000 UTC m=+149.335289242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.024572 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7zklk" podStartSLOduration=129.02455345 podStartE2EDuration="2m9.02455345s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:14.91861094 +0000 UTC m=+148.801237400" watchObservedRunningTime="2025-10-05 06:55:15.02455345 +0000 UTC m=+148.907179910" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.053596 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.054170 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.554154452 +0000 UTC m=+149.436780912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.081752 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-cn4mw" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.145044 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:15 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:15 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:15 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.145659 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.154584 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.155217 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.65519635 +0000 UTC m=+149.537822810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.255871 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.256238 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.756221047 +0000 UTC m=+149.638847507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.352201 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" event={"ID":"cc4744df-9673-48f7-86cb-51df34fcaf17","Type":"ContainerStarted","Data":"8c0e45f8ba7f5cc1c27d36b6af02ee3dde2033789d52994412b75f483d2b8b1e"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.357214 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.358175 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.858159349 +0000 UTC m=+149.740785809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.365581 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-68swt" event={"ID":"ab0901a2-e6ff-44a3-8382-d85f25eeba64","Type":"ContainerStarted","Data":"75c66f6c6388f7266adf90b69fc180a6596495775f4f90f339f47889fbaf945e"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.395546 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-597nr" event={"ID":"22367e16-ef10-498d-8cad-b719508ea6eb","Type":"ContainerStarted","Data":"ff8a0e5bd2dbc724066d903d3ad434c1447f0003362044c15694af37fe2505be"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.395613 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-597nr" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.398594 4935 patch_prober.go:28] interesting pod/downloads-7954f5f757-597nr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.398680 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-597nr" podUID="22367e16-ef10-498d-8cad-b719508ea6eb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.413819 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" event={"ID":"b0bae24b-7c92-4d48-a608-9b7a62c37161","Type":"ContainerStarted","Data":"cfa30c9b1ee57a295c7c7ca2c4ab8cb659e1b64af0a9a82ad02399f7e09e293b"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.419864 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bskzq" event={"ID":"2b54e92f-18aa-4725-a1ec-3b3ef26e6972","Type":"ContainerStarted","Data":"6ea41b537caba753f6122cf107c91cdf7c2eb01befca7ed98a03d45a9da16db3"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.422091 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-597nr" podStartSLOduration=129.42208051 podStartE2EDuration="2m9.42208051s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:15.420777555 +0000 UTC m=+149.303404005" watchObservedRunningTime="2025-10-05 06:55:15.42208051 +0000 UTC m=+149.304706970" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.436388 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" event={"ID":"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7","Type":"ContainerStarted","Data":"d7f20fe688668ac0ffbef28f6104feb0cf02e95e3386ee0240a8534916e8a844"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.448928 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" event={"ID":"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991","Type":"ContainerStarted","Data":"8d7cc13a8ea0f2e82f746e73d29455e0494bf30064b7deb8315eeb5a46737492"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.448970 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" event={"ID":"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991","Type":"ContainerStarted","Data":"a20cb86af93a65ffc01450017aee4ac388c8028622f5ee142d5defd547d35957"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.459937 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.460214 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:15.960202323 +0000 UTC m=+149.842828783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.461398 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" event={"ID":"ac0a9c01-eaff-4da6-a599-48af41b70260","Type":"ContainerStarted","Data":"b846356024c183b133e25c0e7b126e2a97fa759a0567cadab69c332dd1f9e67e"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.480999 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" event={"ID":"ed54eb3b-78b4-4c57-bb72-49d830aede0d","Type":"ContainerStarted","Data":"a9a3ab14a82a761d313832657b595e3386c2043fba84aacb015bb3f6fcc99287"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.486400 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" event={"ID":"e0336e4e-ef2a-42f5-838c-c8e6dd919d59","Type":"ContainerStarted","Data":"67312a156bf9daea9872350e861fc2e2a85674867a606a0b70797e8857ce88bf"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.498810 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" podStartSLOduration=129.498795769 podStartE2EDuration="2m9.498795769s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:15.48555938 +0000 UTC m=+149.368185840" watchObservedRunningTime="2025-10-05 06:55:15.498795769 +0000 UTC m=+149.381422229" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.521735 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" event={"ID":"c55e2cd9-3e55-492c-b005-d30c278bfb03","Type":"ContainerStarted","Data":"83b01fc161cb8a04db6b3b1cb416a46861ff752c560cf46e9c56ac2ab231a881"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.524954 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" event={"ID":"95dcac01-54e5-4fcd-a4eb-068edbe7da4f","Type":"ContainerStarted","Data":"c63501f154fdd7d92bb230d51540306d09be2404bde9c5e21436bffc000c865c"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.530134 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" event={"ID":"4a6790bc-0770-4667-94d9-4cc203809743","Type":"ContainerStarted","Data":"50072021a5fe8c8b24bf8273c3e939f0cac9aa009bda2318d4db1a6e2b019a10"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.535731 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" event={"ID":"c399d3f0-061d-49b7-beb8-3bc3dc378b39","Type":"ContainerStarted","Data":"3e58559ddc958c9b7637721baac181195e2245dfd33bb61a5f32fd83688b9f55"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.549419 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" event={"ID":"c399d3f0-061d-49b7-beb8-3bc3dc378b39","Type":"ContainerStarted","Data":"009a15b1e7d3d56a140fe91577b0fcc0855dbb741b8c9fd779b1a7beb2f9f554"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.549559 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.570923 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.571420 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.071361895 +0000 UTC m=+149.953988355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.572741 4935 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qzm8l container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.572787 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" podUID="c399d3f0-061d-49b7-beb8-3bc3dc378b39" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.575489 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-drgxn" podStartSLOduration=129.575469616 podStartE2EDuration="2m9.575469616s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:15.573105132 +0000 UTC m=+149.455731582" watchObservedRunningTime="2025-10-05 06:55:15.575469616 +0000 UTC m=+149.458096076" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.581353 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.588480 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.088464048 +0000 UTC m=+149.971090508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.648825 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" event={"ID":"7ddea8a8-fd20-4eb2-8321-98f52f5847da","Type":"ContainerStarted","Data":"571967514ad02903bad31648aa6b7f250e14b657f762919b2498f25b9fb10c71"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.678091 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" event={"ID":"23018ad9-0e58-4976-a215-fae5b1b51da6","Type":"ContainerStarted","Data":"d47bf2bcb59f045fadb172b118dd825128cf61c9749b26c5067f6bd88a6cbf20"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.700735 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.701036 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.201018668 +0000 UTC m=+150.083645128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.705010 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" event={"ID":"8df3f651-d6cd-48a2-a97e-84448ce9fb00","Type":"ContainerStarted","Data":"f20f45d05e81d9a3acb9d778e572eba8ed0c4dad2f89bcd4015b040878d6fa50"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.705667 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.714806 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-n2pnb" podStartSLOduration=129.714791331 podStartE2EDuration="2m9.714791331s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:15.714204995 +0000 UTC m=+149.596831455" watchObservedRunningTime="2025-10-05 06:55:15.714791331 +0000 UTC m=+149.597417791" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.749771 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" event={"ID":"c2d4e7a8-02e7-40b1-a024-cfe05f311c17","Type":"ContainerStarted","Data":"851dabb1eb19f25288159ecb21bea0cdfe4ddd3fdb5380c34a35475a16c93d0c"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.802706 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.804868 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.304854941 +0000 UTC m=+150.187481401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.835558 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" podStartSLOduration=129.835543482 podStartE2EDuration="2m9.835543482s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:15.834565286 +0000 UTC m=+149.717191746" watchObservedRunningTime="2025-10-05 06:55:15.835543482 +0000 UTC m=+149.718169942" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.843858 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" event={"ID":"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70","Type":"ContainerStarted","Data":"058088ac0001c69fd8ff06f114cab2dbab3d6b20b713f2f9a12974740cc03fc4"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.854837 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" event={"ID":"8f91ddae-6adc-49a0-ac67-dd37606d3427","Type":"ContainerStarted","Data":"83d5b077a19349ffa33d088d852b4d42b7774d48fce3386516ff3fe0284c68ca"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.856293 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.857965 4935 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fsczh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.858031 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" podUID="8f91ddae-6adc-49a0-ac67-dd37606d3427" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.905413 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:15 crc kubenswrapper[4935]: E1005 06:55:15.905991 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.40594664 +0000 UTC m=+150.288573100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.933717 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sfw52" podStartSLOduration=129.933697672 podStartE2EDuration="2m9.933697672s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:15.933208809 +0000 UTC m=+149.815835269" watchObservedRunningTime="2025-10-05 06:55:15.933697672 +0000 UTC m=+149.816324132" Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.959529 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zdg8q" event={"ID":"6a92699e-19a9-409a-b5e8-b79fc23b5d1b","Type":"ContainerStarted","Data":"fc9aa66d4b2e70a25279963445731de93ff5fe584eb016ff8a1aa67077c6be58"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.977457 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" event={"ID":"7039dd6e-0052-4ea2-8a08-be11c457726d","Type":"ContainerStarted","Data":"73a27cee25f02c1396be6283f43d6895cc5ab018a5478b55dfae6aab807b8c14"} Oct 05 06:55:15 crc kubenswrapper[4935]: I1005 06:55:15.998497 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" event={"ID":"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06","Type":"ContainerStarted","Data":"2e96d9e67f72f927e99295ba12936a39d500887080a8fc881d27f8c38404da97"} Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.008681 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.009449 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.509435674 +0000 UTC m=+150.392062134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.008589 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" event={"ID":"c1af3cc3-3fad-473d-a193-7de470b5ff7c","Type":"ContainerStarted","Data":"f832d555c126815cc5bfd1eca7798b79c2401fc015a307ca8dc23f9553674865"} Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.030330 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" event={"ID":"c33e835e-eefc-4c71-84ba-a7e6784b17bd","Type":"ContainerStarted","Data":"b762c41beecc32f77e971710c69e9211c16ed59cdfcb434e1215af3d20f2ed75"} Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.030384 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" event={"ID":"c33e835e-eefc-4c71-84ba-a7e6784b17bd","Type":"ContainerStarted","Data":"ccd596ae5fc5a3debe10df698a8ef32137dcd088c3721b8dd00e441d4d60678c"} Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.059143 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rxlb" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.097754 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" podStartSLOduration=130.097735816 podStartE2EDuration="2m10.097735816s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.068433702 +0000 UTC m=+149.951060162" watchObservedRunningTime="2025-10-05 06:55:16.097735816 +0000 UTC m=+149.980362276" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.101163 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-468br" podStartSLOduration=130.101155199 podStartE2EDuration="2m10.101155199s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.097458929 +0000 UTC m=+149.980085399" watchObservedRunningTime="2025-10-05 06:55:16.101155199 +0000 UTC m=+149.983781659" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.114118 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.115561 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.615541559 +0000 UTC m=+150.498168019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.169940 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" podStartSLOduration=130.169873551 podStartE2EDuration="2m10.169873551s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.152988513 +0000 UTC m=+150.035614983" watchObservedRunningTime="2025-10-05 06:55:16.169873551 +0000 UTC m=+150.052500011" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.175849 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:16 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:16 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:16 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.176051 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.226404 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.226923 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.726884835 +0000 UTC m=+150.609511295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.274804 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" podStartSLOduration=130.274790403 podStartE2EDuration="2m10.274790403s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.273424256 +0000 UTC m=+150.156050716" watchObservedRunningTime="2025-10-05 06:55:16.274790403 +0000 UTC m=+150.157416863" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.333422 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.333760 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.83374624 +0000 UTC m=+150.716372690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.361696 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" podStartSLOduration=130.361669297 podStartE2EDuration="2m10.361669297s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.320780039 +0000 UTC m=+150.203406499" watchObservedRunningTime="2025-10-05 06:55:16.361669297 +0000 UTC m=+150.244295757" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.364922 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-zdg8q" podStartSLOduration=7.364898564 podStartE2EDuration="7.364898564s" podCreationTimestamp="2025-10-05 06:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.360508936 +0000 UTC m=+150.243135396" watchObservedRunningTime="2025-10-05 06:55:16.364898564 +0000 UTC m=+150.247525014" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.435583 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.436073 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:16.936055662 +0000 UTC m=+150.818682122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.467576 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ckxt" podStartSLOduration=130.467556086 podStartE2EDuration="2m10.467556086s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.412791042 +0000 UTC m=+150.295417502" watchObservedRunningTime="2025-10-05 06:55:16.467556086 +0000 UTC m=+150.350182536" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.539430 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.540313 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.040290787 +0000 UTC m=+150.922917247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.563918 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zp4tc" podStartSLOduration=130.563875095 podStartE2EDuration="2m10.563875095s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.468528482 +0000 UTC m=+150.351154942" watchObservedRunningTime="2025-10-05 06:55:16.563875095 +0000 UTC m=+150.446501555" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.609306 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.609723 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.642360 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.642880 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.142860275 +0000 UTC m=+151.025486735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.745965 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.746754 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.246687648 +0000 UTC m=+151.129314118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.775549 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.776300 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.27627647 +0000 UTC m=+151.158902930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.815798 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-579sk" podStartSLOduration=130.815772969 podStartE2EDuration="2m10.815772969s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:16.564593325 +0000 UTC m=+150.447219775" watchObservedRunningTime="2025-10-05 06:55:16.815772969 +0000 UTC m=+150.698399429" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.881589 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.881846 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.381829819 +0000 UTC m=+151.264456279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.943435 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:16 crc kubenswrapper[4935]: I1005 06:55:16.991720 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:16 crc kubenswrapper[4935]: E1005 06:55:16.992283 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.492267321 +0000 UTC m=+151.374893781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.075883 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" event={"ID":"cc4744df-9673-48f7-86cb-51df34fcaf17","Type":"ContainerStarted","Data":"140fbf2ad132e8048d672d450dcc69a368c90cd2d96ef6bfbfc275d0e79855e7"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.096021 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.096323 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.59630457 +0000 UTC m=+151.478931030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.106004 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" event={"ID":"8f91ddae-6adc-49a0-ac67-dd37606d3427","Type":"ContainerStarted","Data":"a75d37202032fe590a341aa4072eb2543fbdd502fea0408ef252f160908867eb"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.107002 4935 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fsczh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.107090 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" podUID="8f91ddae-6adc-49a0-ac67-dd37606d3427" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.145780 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:17 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:17 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:17 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.146239 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.158346 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" event={"ID":"95dcac01-54e5-4fcd-a4eb-068edbe7da4f","Type":"ContainerStarted","Data":"e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.159934 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.162419 4935 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cvn45 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.162465 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" podUID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.191221 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" event={"ID":"7ddea8a8-fd20-4eb2-8321-98f52f5847da","Type":"ContainerStarted","Data":"d5202972e656dfcef22ed115c49f3252a8e119dd3933de3e47c09a7a6f3a3f94"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.198687 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.200496 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.700478852 +0000 UTC m=+151.583105312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.233523 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bskzq" event={"ID":"2b54e92f-18aa-4725-a1ec-3b3ef26e6972","Type":"ContainerStarted","Data":"d217fac495071c2ae6486dab2a9f72a952bfe8dc3e0c71180379ff144ed2abc6"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.234195 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.238790 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" event={"ID":"e0336e4e-ef2a-42f5-838c-c8e6dd919d59","Type":"ContainerStarted","Data":"ef7ef4eacc1ba554fb036399025276643a7bf6fad1c8e7429ca016a30a66bdab"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.242629 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" event={"ID":"4a6790bc-0770-4667-94d9-4cc203809743","Type":"ContainerStarted","Data":"5d74c92cc8ec2858d921ad58e16b981fe3e0762448021b928e80777dd7eaa4b1"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.257313 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wcrn7" event={"ID":"42b7b98d-c2b9-4a61-aabc-e6ec081a3d06","Type":"ContainerStarted","Data":"582e39177ab78614a0d4886480efd32bd7c7e62ebd39b35aa0d8e27afe0a700b"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.275237 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" event={"ID":"ee759263-4b6c-4ee5-9eaa-8de4ed4b9a70","Type":"ContainerStarted","Data":"634ef797435986a49d7834db4ae505544debe639c515b57c0f1e1cc19919c741"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.285494 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-68swt" event={"ID":"ab0901a2-e6ff-44a3-8382-d85f25eeba64","Type":"ContainerStarted","Data":"06345bc1d85cfc4f8dbd9a2ad13c80ca7ed82eb4ffbd2f784810614be467f8fa"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.288826 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7a372497b43b97fa77275512929ed7649b51f306d4d0c090b262a825ddd64f74"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.289183 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.292368 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"30c321f6696a6997bdef0cbe8206ab17d4abcd1ebf4d513321f9a748b2c5e319"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.292396 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8027f3d838d8dbf028fdf37a3a2fdb9b39cc169ee640a5080bbb72dc2f9a1cd7"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.300386 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.301625 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.801597742 +0000 UTC m=+151.684224202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.302422 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.303073 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.803064682 +0000 UTC m=+151.685691142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.324647 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f254r" event={"ID":"fb8e8feb-a6f4-4ace-a1da-a603f41feb41","Type":"ContainerStarted","Data":"e95a7797dbaf79717920e4f78f53f38939caddee787f08584dc5501f0108409c"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.347086 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" event={"ID":"b0bae24b-7c92-4d48-a608-9b7a62c37161","Type":"ContainerStarted","Data":"3451b11f56e15889bcb102fdc8d1b908c693e1657dc052bfe613d5c1d5bb8100"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.352998 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"69327b7c51d254b22d7df9f567cd5eba931448d8d176e546c11f1854575d2fc7"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.360717 4935 generic.go:334] "Generic (PLEG): container finished" podID="9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7" containerID="b23bdc41a297e046333ceada8e5f3784fd558809eb4fdbb81023f1aec281f21d" exitCode=0 Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.363084 4935 patch_prober.go:28] interesting pod/downloads-7954f5f757-597nr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.363174 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-597nr" podUID="22367e16-ef10-498d-8cad-b719508ea6eb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.364180 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" event={"ID":"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7","Type":"ContainerDied","Data":"b23bdc41a297e046333ceada8e5f3784fd558809eb4fdbb81023f1aec281f21d"} Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.380123 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rzp8" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.403536 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.405518 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:17.905494157 +0000 UTC m=+151.788120617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.481216 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hkqkc" podStartSLOduration=131.481181277 podStartE2EDuration="2m11.481181277s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:17.479334337 +0000 UTC m=+151.361960797" watchObservedRunningTime="2025-10-05 06:55:17.481181277 +0000 UTC m=+151.363807737" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.505652 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.517667 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.017640575 +0000 UTC m=+151.900267225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.594511 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-trqss" podStartSLOduration=131.594494788 podStartE2EDuration="2m11.594494788s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:17.592977756 +0000 UTC m=+151.475604206" watchObservedRunningTime="2025-10-05 06:55:17.594494788 +0000 UTC m=+151.477121248" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.610950 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.612005 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.111985131 +0000 UTC m=+151.994611591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.674321 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" podStartSLOduration=131.674258919 podStartE2EDuration="2m11.674258919s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:17.657397502 +0000 UTC m=+151.540023982" watchObservedRunningTime="2025-10-05 06:55:17.674258919 +0000 UTC m=+151.556885379" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.713505 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.714398 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.214382506 +0000 UTC m=+152.097008966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.811462 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6tkbm" podStartSLOduration=131.811438905 podStartE2EDuration="2m11.811438905s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:17.741170892 +0000 UTC m=+151.623797352" watchObservedRunningTime="2025-10-05 06:55:17.811438905 +0000 UTC m=+151.694065355" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.814788 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.814871 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.314853158 +0000 UTC m=+152.197479608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.824081 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.824613 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.324596942 +0000 UTC m=+152.207223402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.837808 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mzckv" podStartSLOduration=131.837783519 podStartE2EDuration="2m11.837783519s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:17.837241344 +0000 UTC m=+151.719867794" watchObservedRunningTime="2025-10-05 06:55:17.837783519 +0000 UTC m=+151.720409969" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.841137 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qzm8l" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.897559 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-sbwkd" podStartSLOduration=131.897531908 podStartE2EDuration="2m11.897531908s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:17.893748815 +0000 UTC m=+151.776375285" watchObservedRunningTime="2025-10-05 06:55:17.897531908 +0000 UTC m=+151.780158368" Oct 05 06:55:17 crc kubenswrapper[4935]: I1005 06:55:17.926705 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:17 crc kubenswrapper[4935]: E1005 06:55:17.927449 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.427427768 +0000 UTC m=+152.310054228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.032691 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.033216 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.533188693 +0000 UTC m=+152.415815153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.033536 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-f254r" podStartSLOduration=132.033511392 podStartE2EDuration="2m12.033511392s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:18.031425956 +0000 UTC m=+151.914052446" watchObservedRunningTime="2025-10-05 06:55:18.033511392 +0000 UTC m=+151.916137852" Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.060021 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-68swt" podStartSLOduration=9.06000604 podStartE2EDuration="9.06000604s" podCreationTimestamp="2025-10-05 06:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:18.059916948 +0000 UTC m=+151.942543418" watchObservedRunningTime="2025-10-05 06:55:18.06000604 +0000 UTC m=+151.942632500" Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.133379 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.133783 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.633757878 +0000 UTC m=+152.516384338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.150278 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:18 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:18 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:18 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.150824 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.168285 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-bskzq" podStartSLOduration=9.168267943 podStartE2EDuration="9.168267943s" podCreationTimestamp="2025-10-05 06:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:18.12201516 +0000 UTC m=+152.004641620" watchObservedRunningTime="2025-10-05 06:55:18.168267943 +0000 UTC m=+152.050894403" Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.236205 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.236983 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.736962844 +0000 UTC m=+152.619589294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.338480 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.338632 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.838612278 +0000 UTC m=+152.721238728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.338728 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.339091 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.839083531 +0000 UTC m=+152.721709991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.368710 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c90f74e1a0a9b9a4e7df2bcc2f8c8714564c2592399d670916045433de6477c6"} Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.371587 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f254r" event={"ID":"fb8e8feb-a6f4-4ace-a1da-a603f41feb41","Type":"ContainerStarted","Data":"c6bf02d5d4ae9a1e4ba6c399a82cc2ba5ab49d06252fdadc1a1956e74952ab1a"} Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.373934 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bskzq" event={"ID":"2b54e92f-18aa-4725-a1ec-3b3ef26e6972","Type":"ContainerStarted","Data":"212a29d70be2b9d8714d9ffa03373274df6b89874568b6e80cde1a7e97b84755"} Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.376146 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" event={"ID":"9fa12988-6e6d-4a16-9cca-3b96a4e4fbb7","Type":"ContainerStarted","Data":"f883d94165fb516cbb8a439ec3f9cdff0e16c348c55057ae617e21a8da35b5ee"} Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.376361 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.377751 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1a3606ffbb09f3aeed578729155d7eb9aa216dc78e7623f370fcc62fd638f17c"} Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.378235 4935 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cvn45 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.378285 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" podUID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.398470 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fsczh" Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.412571 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" podStartSLOduration=132.412552792 podStartE2EDuration="2m12.412552792s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:18.410474055 +0000 UTC m=+152.293100535" watchObservedRunningTime="2025-10-05 06:55:18.412552792 +0000 UTC m=+152.295179252" Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.440144 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.440329 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.940298893 +0000 UTC m=+152.822925353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.440513 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.441551 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:18.941534687 +0000 UTC m=+152.824161237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.541533 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.542671 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.042646826 +0000 UTC m=+152.925273286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.649169 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.649462 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.14945023 +0000 UTC m=+153.032076690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.750665 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.751339 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.25132 +0000 UTC m=+153.133946460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.852084 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.852495 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.352478031 +0000 UTC m=+153.235104491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:18 crc kubenswrapper[4935]: I1005 06:55:18.953559 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:18 crc kubenswrapper[4935]: E1005 06:55:18.953854 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.453836957 +0000 UTC m=+153.336463417 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.054786 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.055220 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.555203773 +0000 UTC m=+153.437830233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.080756 4935 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.147935 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:19 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:19 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:19 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.148972 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.156440 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.156910 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.656861858 +0000 UTC m=+153.539488318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.218210 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p4q5n"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.219195 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.222224 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.232593 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4q5n"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.258468 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.258795 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.758782699 +0000 UTC m=+153.641409159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.358624 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.359248 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.359375 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.359429 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.859403975 +0000 UTC m=+153.742030425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.359464 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-utilities\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.359511 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqkmx\" (UniqueName: \"kubernetes.io/projected/a1476c09-cca4-4cd9-adfd-188e00dc9457-kube-api-access-tqkmx\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.359554 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-catalog-content\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.359661 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.360013 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.859998351 +0000 UTC m=+153.742624801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.362631 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.364270 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.377000 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.385748 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" event={"ID":"e0336e4e-ef2a-42f5-838c-c8e6dd919d59","Type":"ContainerStarted","Data":"3a771ead94c28aac403a01e3de5e129efd7f01ab91029b517f0e69f698b131c8"} Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.386004 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" event={"ID":"e0336e4e-ef2a-42f5-838c-c8e6dd919d59","Type":"ContainerStarted","Data":"3d35965ac9689792b18c7342bae0482427339e33d6c2f0d0227800f894a04136"} Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.386078 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" event={"ID":"e0336e4e-ef2a-42f5-838c-c8e6dd919d59","Type":"ContainerStarted","Data":"c873c9a7af8cc992de4832964f3e1e2f6c8c4c7cadc09d2ae719e41cb5a11e4d"} Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.387349 4935 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cvn45 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.387424 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" podUID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.413350 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6rk8k"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.414438 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.416121 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.433803 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-ssjkm" podStartSLOduration=10.43378339 podStartE2EDuration="10.43378339s" podCreationTimestamp="2025-10-05 06:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:19.431709114 +0000 UTC m=+153.314335574" watchObservedRunningTime="2025-10-05 06:55:19.43378339 +0000 UTC m=+153.316409850" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.443337 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6rk8k"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.462233 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.462678 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f03a883-a9c1-414d-9d3f-b9930273821c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1f03a883-a9c1-414d-9d3f-b9930273821c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.462813 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqkmx\" (UniqueName: \"kubernetes.io/projected/a1476c09-cca4-4cd9-adfd-188e00dc9457-kube-api-access-tqkmx\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.462982 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f03a883-a9c1-414d-9d3f-b9930273821c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1f03a883-a9c1-414d-9d3f-b9930273821c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.463032 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-catalog-content\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.463364 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-utilities\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.463862 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-utilities\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.463911 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:19.963843175 +0000 UTC m=+153.846469645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.474726 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-catalog-content\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.504279 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqkmx\" (UniqueName: \"kubernetes.io/projected/a1476c09-cca4-4cd9-adfd-188e00dc9457-kube-api-access-tqkmx\") pod \"community-operators-p4q5n\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.547728 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.564524 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-utilities\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.564593 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f03a883-a9c1-414d-9d3f-b9930273821c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1f03a883-a9c1-414d-9d3f-b9930273821c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.564688 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v4g9\" (UniqueName: \"kubernetes.io/projected/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-kube-api-access-9v4g9\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.564715 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f03a883-a9c1-414d-9d3f-b9930273821c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1f03a883-a9c1-414d-9d3f-b9930273821c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.564757 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.564792 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-catalog-content\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.565227 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f03a883-a9c1-414d-9d3f-b9930273821c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1f03a883-a9c1-414d-9d3f-b9930273821c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.565526 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:20.065510479 +0000 UTC m=+153.948136939 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.600788 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f03a883-a9c1-414d-9d3f-b9930273821c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1f03a883-a9c1-414d-9d3f-b9930273821c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.603373 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wzkhf"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.604342 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.618410 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wzkhf"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.671086 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.671269 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:20.171240274 +0000 UTC m=+154.053866734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.671306 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.671352 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-catalog-content\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.671390 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-utilities\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.671440 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v4g9\" (UniqueName: \"kubernetes.io/projected/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-kube-api-access-9v4g9\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.671615 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:20.171608134 +0000 UTC m=+154.054234594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.671990 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-catalog-content\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.672308 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-utilities\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.674937 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.692704 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v4g9\" (UniqueName: \"kubernetes.io/projected/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-kube-api-access-9v4g9\") pod \"certified-operators-6rk8k\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.727471 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.776199 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.776472 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-catalog-content\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.776512 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htqlq\" (UniqueName: \"kubernetes.io/projected/30c3e073-685a-4fe1-bdb3-12289e64075b-kube-api-access-htqlq\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.776557 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-utilities\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.776695 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:55:20.276679491 +0000 UTC m=+154.159305951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.805548 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7kqbg"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.806640 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.825075 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7kqbg"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.878333 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-utilities\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.878392 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.878435 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-catalog-content\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.878464 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htqlq\" (UniqueName: \"kubernetes.io/projected/30c3e073-685a-4fe1-bdb3-12289e64075b-kube-api-access-htqlq\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.885049 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-utilities\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: E1005 06:55:19.885365 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:55:20.385352925 +0000 UTC m=+154.267979375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5skck" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.885736 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-catalog-content\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.892077 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4q5n"] Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.896607 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htqlq\" (UniqueName: \"kubernetes.io/projected/30c3e073-685a-4fe1-bdb3-12289e64075b-kube-api-access-htqlq\") pod \"community-operators-wzkhf\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.968522 4935 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-05T06:55:19.081181627Z","Handler":null,"Name":""} Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.972303 4935 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.972329 4935 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.979083 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.979330 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-catalog-content\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.979429 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24fts\" (UniqueName: \"kubernetes.io/projected/6f8beac5-0f55-450a-8d18-d4f331076e26-kube-api-access-24fts\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.979472 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-utilities\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.980243 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:55:19 crc kubenswrapper[4935]: I1005 06:55:19.994436 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.065761 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.071791 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6rk8k"] Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.084527 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24fts\" (UniqueName: \"kubernetes.io/projected/6f8beac5-0f55-450a-8d18-d4f331076e26-kube-api-access-24fts\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.084597 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-utilities\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.085177 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-utilities\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.085214 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.085264 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-catalog-content\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.085626 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-catalog-content\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.124079 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24fts\" (UniqueName: \"kubernetes.io/projected/6f8beac5-0f55-450a-8d18-d4f331076e26-kube-api-access-24fts\") pod \"certified-operators-7kqbg\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.131109 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.144725 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:20 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:20 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:20 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.144785 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.148375 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.148425 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.218812 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5skck\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.231489 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.321858 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wzkhf"] Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.404119 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rk8k" event={"ID":"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6","Type":"ContainerStarted","Data":"d2238137491d4f91f2ab0a759e3f49cba805a9bef9d67a461d504976b6a5898c"} Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.415886 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1f03a883-a9c1-414d-9d3f-b9930273821c","Type":"ContainerStarted","Data":"45ad4bbde1b1ac30c89ec70463dd9ca5b917646297421300ad205549d80f560c"} Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.421846 4935 generic.go:334] "Generic (PLEG): container finished" podID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerID="5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d" exitCode=0 Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.422114 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q5n" event={"ID":"a1476c09-cca4-4cd9-adfd-188e00dc9457","Type":"ContainerDied","Data":"5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d"} Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.422138 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q5n" event={"ID":"a1476c09-cca4-4cd9-adfd-188e00dc9457","Type":"ContainerStarted","Data":"c1fd3b37fbc002f3729303639b41030c5849544ee4974563c17d059b6e2d9c27"} Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.423349 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.423618 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzkhf" event={"ID":"30c3e073-685a-4fe1-bdb3-12289e64075b","Type":"ContainerStarted","Data":"f5d74bba98044d41cd81fff7e1f0a2f7b815b3b65841f8fde0d4768dc00921b4"} Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.478408 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7kqbg"] Oct 05 06:55:20 crc kubenswrapper[4935]: W1005 06:55:20.491644 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f8beac5_0f55_450a_8d18_d4f331076e26.slice/crio-c2c9f31e88e9d74b760d875cccb178a52a801bf65ae6d2ac831e01403e66d076 WatchSource:0}: Error finding container c2c9f31e88e9d74b760d875cccb178a52a801bf65ae6d2ac831e01403e66d076: Status 404 returned error can't find the container with id c2c9f31e88e9d74b760d875cccb178a52a801bf65ae6d2ac831e01403e66d076 Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.564075 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5skck"] Oct 05 06:55:20 crc kubenswrapper[4935]: W1005 06:55:20.566376 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda34ebce5_75e3_4033_9fa0_bb17b2151fae.slice/crio-ad3e5031b0fd2956c9a8fc76d178d71234eeffb01931b759e0f25fc2d137799d WatchSource:0}: Error finding container ad3e5031b0fd2956c9a8fc76d178d71234eeffb01931b759e0f25fc2d137799d: Status 404 returned error can't find the container with id ad3e5031b0fd2956c9a8fc76d178d71234eeffb01931b759e0f25fc2d137799d Oct 05 06:55:20 crc kubenswrapper[4935]: I1005 06:55:20.789220 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.146004 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:21 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:21 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:21 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.148037 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.404899 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fjxc9"] Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.406031 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.408465 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.428690 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjxc9"] Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.432197 4935 generic.go:334] "Generic (PLEG): container finished" podID="7ce9bae6-1db9-4a69-bef4-1f5da3ea1991" containerID="8d7cc13a8ea0f2e82f746e73d29455e0494bf30064b7deb8315eeb5a46737492" exitCode=0 Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.432278 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" event={"ID":"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991","Type":"ContainerDied","Data":"8d7cc13a8ea0f2e82f746e73d29455e0494bf30064b7deb8315eeb5a46737492"} Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.448399 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" event={"ID":"a34ebce5-75e3-4033-9fa0-bb17b2151fae","Type":"ContainerStarted","Data":"4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734"} Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.448445 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" event={"ID":"a34ebce5-75e3-4033-9fa0-bb17b2151fae","Type":"ContainerStarted","Data":"ad3e5031b0fd2956c9a8fc76d178d71234eeffb01931b759e0f25fc2d137799d"} Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.449063 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.450638 4935 generic.go:334] "Generic (PLEG): container finished" podID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerID="a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a" exitCode=0 Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.450690 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7kqbg" event={"ID":"6f8beac5-0f55-450a-8d18-d4f331076e26","Type":"ContainerDied","Data":"a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a"} Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.450706 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7kqbg" event={"ID":"6f8beac5-0f55-450a-8d18-d4f331076e26","Type":"ContainerStarted","Data":"c2c9f31e88e9d74b760d875cccb178a52a801bf65ae6d2ac831e01403e66d076"} Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.455009 4935 generic.go:334] "Generic (PLEG): container finished" podID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerID="b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac" exitCode=0 Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.455774 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rk8k" event={"ID":"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6","Type":"ContainerDied","Data":"b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac"} Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.459705 4935 generic.go:334] "Generic (PLEG): container finished" podID="1f03a883-a9c1-414d-9d3f-b9930273821c" containerID="60f59202582644ed4e6d9e9a24129e0f9a4d17c719cb7fce38843001011e42f4" exitCode=0 Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.459773 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1f03a883-a9c1-414d-9d3f-b9930273821c","Type":"ContainerDied","Data":"60f59202582644ed4e6d9e9a24129e0f9a4d17c719cb7fce38843001011e42f4"} Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.463920 4935 generic.go:334] "Generic (PLEG): container finished" podID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerID="cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077" exitCode=0 Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.463972 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzkhf" event={"ID":"30c3e073-685a-4fe1-bdb3-12289e64075b","Type":"ContainerDied","Data":"cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077"} Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.504705 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" podStartSLOduration=135.504687978 podStartE2EDuration="2m15.504687978s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:21.502565241 +0000 UTC m=+155.385191701" watchObservedRunningTime="2025-10-05 06:55:21.504687978 +0000 UTC m=+155.387314438" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.518399 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-catalog-content\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.518480 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-utilities\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.518515 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxkwc\" (UniqueName: \"kubernetes.io/projected/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-kube-api-access-zxkwc\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.622296 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxkwc\" (UniqueName: \"kubernetes.io/projected/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-kube-api-access-zxkwc\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.622492 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-catalog-content\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.622586 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-utilities\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.623537 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-utilities\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.623557 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-catalog-content\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.630768 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gx7h8" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.661822 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxkwc\" (UniqueName: \"kubernetes.io/projected/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-kube-api-access-zxkwc\") pod \"redhat-marketplace-fjxc9\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.725340 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.725449 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.725500 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.727092 4935 patch_prober.go:28] interesting pod/console-f9d7485db-cbb4c container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.727135 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-cbb4c" podUID="ac0661f5-9317-459e-bfce-c2ff9058c319" containerName="console" probeResult="failure" output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.730305 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.731426 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.736987 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.802938 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v8lrt"] Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.804152 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.810880 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v8lrt"] Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.929314 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-catalog-content\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.929375 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfzw7\" (UniqueName: \"kubernetes.io/projected/05aae77b-1528-4993-9187-a4b4513fcf4f-kube-api-access-wfzw7\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:21 crc kubenswrapper[4935]: I1005 06:55:21.929439 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-utilities\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.024388 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjxc9"] Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.030960 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-catalog-content\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.031021 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfzw7\" (UniqueName: \"kubernetes.io/projected/05aae77b-1528-4993-9187-a4b4513fcf4f-kube-api-access-wfzw7\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.031054 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-utilities\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.031741 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-utilities\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.032204 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-catalog-content\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:22 crc kubenswrapper[4935]: W1005 06:55:22.036340 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod638ea9d6_e6c3_4ae4_b361_31ca73cb242a.slice/crio-fa83e0c92b23205de8c2257539a8baac2785cb1d666d81c8e365c9a0dc9de35a WatchSource:0}: Error finding container fa83e0c92b23205de8c2257539a8baac2785cb1d666d81c8e365c9a0dc9de35a: Status 404 returned error can't find the container with id fa83e0c92b23205de8c2257539a8baac2785cb1d666d81c8e365c9a0dc9de35a Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.050139 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfzw7\" (UniqueName: \"kubernetes.io/projected/05aae77b-1528-4993-9187-a4b4513fcf4f-kube-api-access-wfzw7\") pod \"redhat-marketplace-v8lrt\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.143163 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.147278 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:22 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:22 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:22 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.147331 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.153111 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.379103 4935 patch_prober.go:28] interesting pod/downloads-7954f5f757-597nr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.379570 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-597nr" podUID="22367e16-ef10-498d-8cad-b719508ea6eb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.379831 4935 patch_prober.go:28] interesting pod/downloads-7954f5f757-597nr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.379848 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-597nr" podUID="22367e16-ef10-498d-8cad-b719508ea6eb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.401812 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8rmsz"] Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.402919 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.416507 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.419843 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8rmsz"] Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.516147 4935 generic.go:334] "Generic (PLEG): container finished" podID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerID="a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8" exitCode=0 Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.517189 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjxc9" event={"ID":"638ea9d6-e6c3-4ae4-b361-31ca73cb242a","Type":"ContainerDied","Data":"a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8"} Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.517296 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjxc9" event={"ID":"638ea9d6-e6c3-4ae4-b361-31ca73cb242a","Type":"ContainerStarted","Data":"fa83e0c92b23205de8c2257539a8baac2785cb1d666d81c8e365c9a0dc9de35a"} Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.524443 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-f254r" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.549950 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-catalog-content\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.550054 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz62k\" (UniqueName: \"kubernetes.io/projected/ce5d3e82-fc79-483f-8e45-8fc712891e7c-kube-api-access-vz62k\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.550073 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-utilities\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.651621 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-catalog-content\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.651810 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz62k\" (UniqueName: \"kubernetes.io/projected/ce5d3e82-fc79-483f-8e45-8fc712891e7c-kube-api-access-vz62k\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.651830 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-utilities\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.652668 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-utilities\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.653066 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-catalog-content\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.684846 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz62k\" (UniqueName: \"kubernetes.io/projected/ce5d3e82-fc79-483f-8e45-8fc712891e7c-kube-api-access-vz62k\") pod \"redhat-operators-8rmsz\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.719824 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v8lrt"] Oct 05 06:55:22 crc kubenswrapper[4935]: W1005 06:55:22.739115 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05aae77b_1528_4993_9187_a4b4513fcf4f.slice/crio-b3a4affb976ac5efd631e0ece66f4e79dfb8292043cb9ec92effd93640b07973 WatchSource:0}: Error finding container b3a4affb976ac5efd631e0ece66f4e79dfb8292043cb9ec92effd93640b07973: Status 404 returned error can't find the container with id b3a4affb976ac5efd631e0ece66f4e79dfb8292043cb9ec92effd93640b07973 Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.744182 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.823420 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4scg9"] Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.842284 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.856719 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4scg9"] Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.951370 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.958264 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj842\" (UniqueName: \"kubernetes.io/projected/70f5830f-6e20-4066-9419-7916d6a6c0ca-kube-api-access-gj842\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.958388 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-catalog-content\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:22 crc kubenswrapper[4935]: I1005 06:55:22.958597 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-utilities\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.059934 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f03a883-a9c1-414d-9d3f-b9930273821c-kubelet-dir\") pod \"1f03a883-a9c1-414d-9d3f-b9930273821c\" (UID: \"1f03a883-a9c1-414d-9d3f-b9930273821c\") " Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.060021 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f03a883-a9c1-414d-9d3f-b9930273821c-kube-api-access\") pod \"1f03a883-a9c1-414d-9d3f-b9930273821c\" (UID: \"1f03a883-a9c1-414d-9d3f-b9930273821c\") " Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.060191 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-utilities\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.060224 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj842\" (UniqueName: \"kubernetes.io/projected/70f5830f-6e20-4066-9419-7916d6a6c0ca-kube-api-access-gj842\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.060280 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-catalog-content\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.061188 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-catalog-content\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.080145 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8rmsz"] Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.080316 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-utilities\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.080340 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f03a883-a9c1-414d-9d3f-b9930273821c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1f03a883-a9c1-414d-9d3f-b9930273821c" (UID: "1f03a883-a9c1-414d-9d3f-b9930273821c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.080996 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f03a883-a9c1-414d-9d3f-b9930273821c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1f03a883-a9c1-414d-9d3f-b9930273821c" (UID: "1f03a883-a9c1-414d-9d3f-b9930273821c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.110311 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj842\" (UniqueName: \"kubernetes.io/projected/70f5830f-6e20-4066-9419-7916d6a6c0ca-kube-api-access-gj842\") pod \"redhat-operators-4scg9\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.134352 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.145410 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:23 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:23 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:23 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.145493 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.162080 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f03a883-a9c1-414d-9d3f-b9930273821c-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.162128 4935 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1f03a883-a9c1-414d-9d3f-b9930273821c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.173880 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.243004 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.262935 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume\") pod \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.263025 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxpzv\" (UniqueName: \"kubernetes.io/projected/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-kube-api-access-kxpzv\") pod \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.263094 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-secret-volume\") pod \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\" (UID: \"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991\") " Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.263836 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume" (OuterVolumeSpecName: "config-volume") pod "7ce9bae6-1db9-4a69-bef4-1f5da3ea1991" (UID: "7ce9bae6-1db9-4a69-bef4-1f5da3ea1991"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.267788 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-kube-api-access-kxpzv" (OuterVolumeSpecName: "kube-api-access-kxpzv") pod "7ce9bae6-1db9-4a69-bef4-1f5da3ea1991" (UID: "7ce9bae6-1db9-4a69-bef4-1f5da3ea1991"). InnerVolumeSpecName "kube-api-access-kxpzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.268579 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7ce9bae6-1db9-4a69-bef4-1f5da3ea1991" (UID: "7ce9bae6-1db9-4a69-bef4-1f5da3ea1991"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.364972 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.365015 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.365029 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxpzv\" (UniqueName: \"kubernetes.io/projected/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991-kube-api-access-kxpzv\") on node \"crc\" DevicePath \"\"" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.507785 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4scg9"] Oct 05 06:55:23 crc kubenswrapper[4935]: W1005 06:55:23.527011 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70f5830f_6e20_4066_9419_7916d6a6c0ca.slice/crio-ba08060bc3b3d1d9a83c8084c67a52ab5cbf9476a0ef56cf88db24bfed0e6a47 WatchSource:0}: Error finding container ba08060bc3b3d1d9a83c8084c67a52ab5cbf9476a0ef56cf88db24bfed0e6a47: Status 404 returned error can't find the container with id ba08060bc3b3d1d9a83c8084c67a52ab5cbf9476a0ef56cf88db24bfed0e6a47 Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.533202 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1f03a883-a9c1-414d-9d3f-b9930273821c","Type":"ContainerDied","Data":"45ad4bbde1b1ac30c89ec70463dd9ca5b917646297421300ad205549d80f560c"} Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.533660 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45ad4bbde1b1ac30c89ec70463dd9ca5b917646297421300ad205549d80f560c" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.533784 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.539657 4935 generic.go:334] "Generic (PLEG): container finished" podID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerID="f3159b808d14b5719191980922c6a426c7f7a1659da952a3d14c4d31b4f0a43b" exitCode=0 Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.539776 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8lrt" event={"ID":"05aae77b-1528-4993-9187-a4b4513fcf4f","Type":"ContainerDied","Data":"f3159b808d14b5719191980922c6a426c7f7a1659da952a3d14c4d31b4f0a43b"} Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.539825 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8lrt" event={"ID":"05aae77b-1528-4993-9187-a4b4513fcf4f","Type":"ContainerStarted","Data":"b3a4affb976ac5efd631e0ece66f4e79dfb8292043cb9ec92effd93640b07973"} Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.547155 4935 generic.go:334] "Generic (PLEG): container finished" podID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerID="2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748" exitCode=0 Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.547237 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rmsz" event={"ID":"ce5d3e82-fc79-483f-8e45-8fc712891e7c","Type":"ContainerDied","Data":"2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748"} Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.547267 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rmsz" event={"ID":"ce5d3e82-fc79-483f-8e45-8fc712891e7c","Type":"ContainerStarted","Data":"808b4260bdd958f059421fd47ea50d6b82b5548dac188c2073c31ddfa3befbd1"} Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.560340 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" event={"ID":"7ce9bae6-1db9-4a69-bef4-1f5da3ea1991","Type":"ContainerDied","Data":"a20cb86af93a65ffc01450017aee4ac388c8028622f5ee142d5defd547d35957"} Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.560379 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a20cb86af93a65ffc01450017aee4ac388c8028622f5ee142d5defd547d35957" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.560416 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.618269 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 06:55:23 crc kubenswrapper[4935]: E1005 06:55:23.618492 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce9bae6-1db9-4a69-bef4-1f5da3ea1991" containerName="collect-profiles" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.618504 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce9bae6-1db9-4a69-bef4-1f5da3ea1991" containerName="collect-profiles" Oct 05 06:55:23 crc kubenswrapper[4935]: E1005 06:55:23.618523 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f03a883-a9c1-414d-9d3f-b9930273821c" containerName="pruner" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.618529 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f03a883-a9c1-414d-9d3f-b9930273821c" containerName="pruner" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.618677 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce9bae6-1db9-4a69-bef4-1f5da3ea1991" containerName="collect-profiles" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.618694 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f03a883-a9c1-414d-9d3f-b9930273821c" containerName="pruner" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.619255 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.621613 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.621792 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.628225 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.771327 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba77708b-bb6a-4b80-bfa2-03f36b142595-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ba77708b-bb6a-4b80-bfa2-03f36b142595\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.771419 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba77708b-bb6a-4b80-bfa2-03f36b142595-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ba77708b-bb6a-4b80-bfa2-03f36b142595\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.873198 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba77708b-bb6a-4b80-bfa2-03f36b142595-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ba77708b-bb6a-4b80-bfa2-03f36b142595\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.873260 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba77708b-bb6a-4b80-bfa2-03f36b142595-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ba77708b-bb6a-4b80-bfa2-03f36b142595\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.873347 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba77708b-bb6a-4b80-bfa2-03f36b142595-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ba77708b-bb6a-4b80-bfa2-03f36b142595\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.893046 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba77708b-bb6a-4b80-bfa2-03f36b142595-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ba77708b-bb6a-4b80-bfa2-03f36b142595\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:23 crc kubenswrapper[4935]: I1005 06:55:23.948121 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:24 crc kubenswrapper[4935]: I1005 06:55:24.145570 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:24 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:24 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:24 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:24 crc kubenswrapper[4935]: I1005 06:55:24.146041 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:24 crc kubenswrapper[4935]: I1005 06:55:24.310327 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 06:55:24 crc kubenswrapper[4935]: I1005 06:55:24.584977 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ba77708b-bb6a-4b80-bfa2-03f36b142595","Type":"ContainerStarted","Data":"0827750d802b8736598a78648c1c035744a0f894b04dd92ed0dcef7080f0d7ad"} Oct 05 06:55:24 crc kubenswrapper[4935]: I1005 06:55:24.588553 4935 generic.go:334] "Generic (PLEG): container finished" podID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerID="729342ad817013847f0be059797677d7236d3eee5193aa794056416d1c60ce2e" exitCode=0 Oct 05 06:55:24 crc kubenswrapper[4935]: I1005 06:55:24.588585 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4scg9" event={"ID":"70f5830f-6e20-4066-9419-7916d6a6c0ca","Type":"ContainerDied","Data":"729342ad817013847f0be059797677d7236d3eee5193aa794056416d1c60ce2e"} Oct 05 06:55:24 crc kubenswrapper[4935]: I1005 06:55:24.588604 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4scg9" event={"ID":"70f5830f-6e20-4066-9419-7916d6a6c0ca","Type":"ContainerStarted","Data":"ba08060bc3b3d1d9a83c8084c67a52ab5cbf9476a0ef56cf88db24bfed0e6a47"} Oct 05 06:55:25 crc kubenswrapper[4935]: I1005 06:55:25.143479 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:25 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:25 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:25 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:25 crc kubenswrapper[4935]: I1005 06:55:25.145701 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:25 crc kubenswrapper[4935]: I1005 06:55:25.607203 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ba77708b-bb6a-4b80-bfa2-03f36b142595","Type":"ContainerStarted","Data":"499ba11f7ec926578f3a21734baa827abdc5495d8075cb404d2866d1a3f4200f"} Oct 05 06:55:25 crc kubenswrapper[4935]: I1005 06:55:25.624561 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.62454646 podStartE2EDuration="2.62454646s" podCreationTimestamp="2025-10-05 06:55:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:25.620758537 +0000 UTC m=+159.503385007" watchObservedRunningTime="2025-10-05 06:55:25.62454646 +0000 UTC m=+159.507172920" Oct 05 06:55:26 crc kubenswrapper[4935]: I1005 06:55:26.142933 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:26 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:26 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:26 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:26 crc kubenswrapper[4935]: I1005 06:55:26.143006 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:26 crc kubenswrapper[4935]: I1005 06:55:26.618921 4935 generic.go:334] "Generic (PLEG): container finished" podID="ba77708b-bb6a-4b80-bfa2-03f36b142595" containerID="499ba11f7ec926578f3a21734baa827abdc5495d8075cb404d2866d1a3f4200f" exitCode=0 Oct 05 06:55:26 crc kubenswrapper[4935]: I1005 06:55:26.618974 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ba77708b-bb6a-4b80-bfa2-03f36b142595","Type":"ContainerDied","Data":"499ba11f7ec926578f3a21734baa827abdc5495d8075cb404d2866d1a3f4200f"} Oct 05 06:55:27 crc kubenswrapper[4935]: I1005 06:55:27.142862 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:27 crc kubenswrapper[4935]: [-]has-synced failed: reason withheld Oct 05 06:55:27 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:27 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:27 crc kubenswrapper[4935]: I1005 06:55:27.143029 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:27 crc kubenswrapper[4935]: I1005 06:55:27.894652 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-bskzq" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:27.999706 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.143946 4935 patch_prober.go:28] interesting pod/router-default-5444994796-gc2gr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:55:28 crc kubenswrapper[4935]: [+]has-synced ok Oct 05 06:55:28 crc kubenswrapper[4935]: [+]process-running ok Oct 05 06:55:28 crc kubenswrapper[4935]: healthz check failed Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.144024 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc2gr" podUID="8ef91e14-6bf7-47de-b6da-ec630e189389" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.156765 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba77708b-bb6a-4b80-bfa2-03f36b142595-kube-api-access\") pod \"ba77708b-bb6a-4b80-bfa2-03f36b142595\" (UID: \"ba77708b-bb6a-4b80-bfa2-03f36b142595\") " Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.156989 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba77708b-bb6a-4b80-bfa2-03f36b142595-kubelet-dir\") pod \"ba77708b-bb6a-4b80-bfa2-03f36b142595\" (UID: \"ba77708b-bb6a-4b80-bfa2-03f36b142595\") " Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.157304 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba77708b-bb6a-4b80-bfa2-03f36b142595-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ba77708b-bb6a-4b80-bfa2-03f36b142595" (UID: "ba77708b-bb6a-4b80-bfa2-03f36b142595"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.164742 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba77708b-bb6a-4b80-bfa2-03f36b142595-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ba77708b-bb6a-4b80-bfa2-03f36b142595" (UID: "ba77708b-bb6a-4b80-bfa2-03f36b142595"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.258793 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba77708b-bb6a-4b80-bfa2-03f36b142595-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.258829 4935 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba77708b-bb6a-4b80-bfa2-03f36b142595-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.635835 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ba77708b-bb6a-4b80-bfa2-03f36b142595","Type":"ContainerDied","Data":"0827750d802b8736598a78648c1c035744a0f894b04dd92ed0dcef7080f0d7ad"} Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.635880 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0827750d802b8736598a78648c1c035744a0f894b04dd92ed0dcef7080f0d7ad" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.635906 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.677551 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.682011 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87a0c84e-d575-4468-af55-c814e305d34d-metrics-certs\") pod \"network-metrics-daemon-85tzm\" (UID: \"87a0c84e-d575-4468-af55-c814e305d34d\") " pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.699653 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-85tzm" Oct 05 06:55:28 crc kubenswrapper[4935]: I1005 06:55:28.946387 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-85tzm"] Oct 05 06:55:28 crc kubenswrapper[4935]: W1005 06:55:28.956678 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87a0c84e_d575_4468_af55_c814e305d34d.slice/crio-a4a519d1212d5f116479d83a4371cab04a0d88607fe104496375150858d39f73 WatchSource:0}: Error finding container a4a519d1212d5f116479d83a4371cab04a0d88607fe104496375150858d39f73: Status 404 returned error can't find the container with id a4a519d1212d5f116479d83a4371cab04a0d88607fe104496375150858d39f73 Oct 05 06:55:29 crc kubenswrapper[4935]: I1005 06:55:29.144423 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:29 crc kubenswrapper[4935]: I1005 06:55:29.147220 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-gc2gr" Oct 05 06:55:29 crc kubenswrapper[4935]: I1005 06:55:29.667386 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-85tzm" event={"ID":"87a0c84e-d575-4468-af55-c814e305d34d","Type":"ContainerStarted","Data":"855b31f75615c016602275bb8016d33d0c2b0d2af2994ee625d111956eb93fc6"} Oct 05 06:55:29 crc kubenswrapper[4935]: I1005 06:55:29.667807 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-85tzm" event={"ID":"87a0c84e-d575-4468-af55-c814e305d34d","Type":"ContainerStarted","Data":"a4a519d1212d5f116479d83a4371cab04a0d88607fe104496375150858d39f73"} Oct 05 06:55:31 crc kubenswrapper[4935]: I1005 06:55:31.725537 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:31 crc kubenswrapper[4935]: I1005 06:55:31.730445 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 06:55:32 crc kubenswrapper[4935]: I1005 06:55:32.385523 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-597nr" Oct 05 06:55:40 crc kubenswrapper[4935]: I1005 06:55:40.243053 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 06:55:44 crc kubenswrapper[4935]: I1005 06:55:44.289268 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:55:44 crc kubenswrapper[4935]: I1005 06:55:44.289932 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:55:52 crc kubenswrapper[4935]: I1005 06:55:52.191755 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r798b" Oct 05 06:55:54 crc kubenswrapper[4935]: I1005 06:55:54.908588 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:55:58 crc kubenswrapper[4935]: E1005 06:55:58.314178 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 05 06:55:58 crc kubenswrapper[4935]: E1005 06:55:58.314541 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-24fts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7kqbg_openshift-marketplace(6f8beac5-0f55-450a-8d18-d4f331076e26): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:55:58 crc kubenswrapper[4935]: E1005 06:55:58.315841 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7kqbg" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" Oct 05 06:56:00 crc kubenswrapper[4935]: E1005 06:56:00.856174 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7kqbg" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" Oct 05 06:56:02 crc kubenswrapper[4935]: E1005 06:56:02.340587 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 05 06:56:02 crc kubenswrapper[4935]: E1005 06:56:02.340778 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tqkmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-p4q5n_openshift-marketplace(a1476c09-cca4-4cd9-adfd-188e00dc9457): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:56:02 crc kubenswrapper[4935]: E1005 06:56:02.342441 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-p4q5n" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.541048 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-p4q5n" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.649873 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.650392 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vz62k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-8rmsz_openshift-marketplace(ce5d3e82-fc79-483f-8e45-8fc712891e7c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.651633 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-8rmsz" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.670109 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.670277 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-htqlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wzkhf_openshift-marketplace(30c3e073-685a-4fe1-bdb3-12289e64075b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.672325 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wzkhf" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.695690 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.695846 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9v4g9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-6rk8k_openshift-marketplace(5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:56:07 crc kubenswrapper[4935]: E1005 06:56:07.697993 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-6rk8k" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.165101 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wzkhf" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.165981 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-8rmsz" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.167854 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-6rk8k" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.186616 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.186759 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gj842,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4scg9_openshift-marketplace(70f5830f-6e20-4066-9419-7916d6a6c0ca): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.187948 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-4scg9" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.231085 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.231293 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wfzw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-v8lrt_openshift-marketplace(05aae77b-1528-4993-9187-a4b4513fcf4f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.232808 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-v8lrt" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.240201 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.240307 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zxkwc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fjxc9_openshift-marketplace(638ea9d6-e6c3-4ae4-b361-31ca73cb242a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.241559 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fjxc9" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" Oct 05 06:56:08 crc kubenswrapper[4935]: I1005 06:56:08.930473 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-85tzm" event={"ID":"87a0c84e-d575-4468-af55-c814e305d34d","Type":"ContainerStarted","Data":"ddfc14f6d89499fa6a867fe34f5debd39312d40ed6b8ef533f4cda41e4ae69e3"} Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.933011 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-v8lrt" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.933067 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fjxc9" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" Oct 05 06:56:08 crc kubenswrapper[4935]: E1005 06:56:08.933126 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-4scg9" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" Oct 05 06:56:08 crc kubenswrapper[4935]: I1005 06:56:08.969820 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-85tzm" podStartSLOduration=182.969797552 podStartE2EDuration="3m2.969797552s" podCreationTimestamp="2025-10-05 06:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:56:08.963916994 +0000 UTC m=+202.846543454" watchObservedRunningTime="2025-10-05 06:56:08.969797552 +0000 UTC m=+202.852424052" Oct 05 06:56:13 crc kubenswrapper[4935]: I1005 06:56:13.967640 4935 generic.go:334] "Generic (PLEG): container finished" podID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerID="50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9" exitCode=0 Oct 05 06:56:13 crc kubenswrapper[4935]: I1005 06:56:13.967724 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7kqbg" event={"ID":"6f8beac5-0f55-450a-8d18-d4f331076e26","Type":"ContainerDied","Data":"50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9"} Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.289751 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.289835 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.289926 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.291002 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.291177 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8" gracePeriod=600 Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.977840 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8" exitCode=0 Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.978343 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8"} Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.978389 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"1f8f2033fd55c6c815714e0c27ced3951dd73a6a07baa92041ec8b26931c0625"} Oct 05 06:56:14 crc kubenswrapper[4935]: I1005 06:56:14.984025 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7kqbg" event={"ID":"6f8beac5-0f55-450a-8d18-d4f331076e26","Type":"ContainerStarted","Data":"a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52"} Oct 05 06:56:15 crc kubenswrapper[4935]: I1005 06:56:15.030458 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7kqbg" podStartSLOduration=3.139616891 podStartE2EDuration="56.030432263s" podCreationTimestamp="2025-10-05 06:55:19 +0000 UTC" firstStartedPulling="2025-10-05 06:55:21.451864157 +0000 UTC m=+155.334490617" lastFinishedPulling="2025-10-05 06:56:14.342679529 +0000 UTC m=+208.225305989" observedRunningTime="2025-10-05 06:56:15.025388079 +0000 UTC m=+208.908014589" watchObservedRunningTime="2025-10-05 06:56:15.030432263 +0000 UTC m=+208.913058753" Oct 05 06:56:20 crc kubenswrapper[4935]: I1005 06:56:20.024402 4935 generic.go:334] "Generic (PLEG): container finished" podID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerID="06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6" exitCode=0 Oct 05 06:56:20 crc kubenswrapper[4935]: I1005 06:56:20.024438 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q5n" event={"ID":"a1476c09-cca4-4cd9-adfd-188e00dc9457","Type":"ContainerDied","Data":"06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6"} Oct 05 06:56:20 crc kubenswrapper[4935]: I1005 06:56:20.132818 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:56:20 crc kubenswrapper[4935]: I1005 06:56:20.132854 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:56:20 crc kubenswrapper[4935]: I1005 06:56:20.304620 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:56:21 crc kubenswrapper[4935]: I1005 06:56:21.035222 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q5n" event={"ID":"a1476c09-cca4-4cd9-adfd-188e00dc9457","Type":"ContainerStarted","Data":"9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252"} Oct 05 06:56:21 crc kubenswrapper[4935]: I1005 06:56:21.037763 4935 generic.go:334] "Generic (PLEG): container finished" podID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerID="d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e" exitCode=0 Oct 05 06:56:21 crc kubenswrapper[4935]: I1005 06:56:21.037834 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzkhf" event={"ID":"30c3e073-685a-4fe1-bdb3-12289e64075b","Type":"ContainerDied","Data":"d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e"} Oct 05 06:56:21 crc kubenswrapper[4935]: I1005 06:56:21.081797 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p4q5n" podStartSLOduration=2.079683136 podStartE2EDuration="1m2.081766939s" podCreationTimestamp="2025-10-05 06:55:19 +0000 UTC" firstStartedPulling="2025-10-05 06:55:20.423112914 +0000 UTC m=+154.305739374" lastFinishedPulling="2025-10-05 06:56:20.425196707 +0000 UTC m=+214.307823177" observedRunningTime="2025-10-05 06:56:21.075389497 +0000 UTC m=+214.958015967" watchObservedRunningTime="2025-10-05 06:56:21.081766939 +0000 UTC m=+214.964393409" Oct 05 06:56:21 crc kubenswrapper[4935]: I1005 06:56:21.116983 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:56:22 crc kubenswrapper[4935]: I1005 06:56:22.053013 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzkhf" event={"ID":"30c3e073-685a-4fe1-bdb3-12289e64075b","Type":"ContainerStarted","Data":"02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe"} Oct 05 06:56:22 crc kubenswrapper[4935]: I1005 06:56:22.086435 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wzkhf" podStartSLOduration=3.10812761 podStartE2EDuration="1m3.086403707s" podCreationTimestamp="2025-10-05 06:55:19 +0000 UTC" firstStartedPulling="2025-10-05 06:55:21.465125946 +0000 UTC m=+155.347752406" lastFinishedPulling="2025-10-05 06:56:21.443402003 +0000 UTC m=+215.326028503" observedRunningTime="2025-10-05 06:56:22.08403299 +0000 UTC m=+215.966659470" watchObservedRunningTime="2025-10-05 06:56:22.086403707 +0000 UTC m=+215.969030177" Oct 05 06:56:23 crc kubenswrapper[4935]: I1005 06:56:23.061269 4935 generic.go:334] "Generic (PLEG): container finished" podID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerID="231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7" exitCode=0 Oct 05 06:56:23 crc kubenswrapper[4935]: I1005 06:56:23.061350 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjxc9" event={"ID":"638ea9d6-e6c3-4ae4-b361-31ca73cb242a","Type":"ContainerDied","Data":"231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7"} Oct 05 06:56:23 crc kubenswrapper[4935]: I1005 06:56:23.405747 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7kqbg"] Oct 05 06:56:23 crc kubenswrapper[4935]: I1005 06:56:23.406259 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7kqbg" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerName="registry-server" containerID="cri-o://a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52" gracePeriod=2 Oct 05 06:56:23 crc kubenswrapper[4935]: I1005 06:56:23.875407 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.068071 4935 generic.go:334] "Generic (PLEG): container finished" podID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerID="cd9b97de52276d24102887bef5e1847e155b70503bfa04296412f02f3e4f1be0" exitCode=0 Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.068152 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8lrt" event={"ID":"05aae77b-1528-4993-9187-a4b4513fcf4f","Type":"ContainerDied","Data":"cd9b97de52276d24102887bef5e1847e155b70503bfa04296412f02f3e4f1be0"} Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.071794 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rmsz" event={"ID":"ce5d3e82-fc79-483f-8e45-8fc712891e7c","Type":"ContainerStarted","Data":"3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64"} Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.075434 4935 generic.go:334] "Generic (PLEG): container finished" podID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerID="a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52" exitCode=0 Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.075477 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7kqbg" event={"ID":"6f8beac5-0f55-450a-8d18-d4f331076e26","Type":"ContainerDied","Data":"a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52"} Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.075495 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7kqbg" event={"ID":"6f8beac5-0f55-450a-8d18-d4f331076e26","Type":"ContainerDied","Data":"c2c9f31e88e9d74b760d875cccb178a52a801bf65ae6d2ac831e01403e66d076"} Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.075512 4935 scope.go:117] "RemoveContainer" containerID="a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.075602 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7kqbg" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.076202 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24fts\" (UniqueName: \"kubernetes.io/projected/6f8beac5-0f55-450a-8d18-d4f331076e26-kube-api-access-24fts\") pod \"6f8beac5-0f55-450a-8d18-d4f331076e26\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.076254 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-catalog-content\") pod \"6f8beac5-0f55-450a-8d18-d4f331076e26\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.076342 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-utilities\") pod \"6f8beac5-0f55-450a-8d18-d4f331076e26\" (UID: \"6f8beac5-0f55-450a-8d18-d4f331076e26\") " Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.077332 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-utilities" (OuterVolumeSpecName: "utilities") pod "6f8beac5-0f55-450a-8d18-d4f331076e26" (UID: "6f8beac5-0f55-450a-8d18-d4f331076e26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.077830 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.079512 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjxc9" event={"ID":"638ea9d6-e6c3-4ae4-b361-31ca73cb242a","Type":"ContainerStarted","Data":"9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce"} Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.081172 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rk8k" event={"ID":"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6","Type":"ContainerStarted","Data":"8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a"} Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.082812 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f8beac5-0f55-450a-8d18-d4f331076e26-kube-api-access-24fts" (OuterVolumeSpecName: "kube-api-access-24fts") pod "6f8beac5-0f55-450a-8d18-d4f331076e26" (UID: "6f8beac5-0f55-450a-8d18-d4f331076e26"). InnerVolumeSpecName "kube-api-access-24fts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.096033 4935 scope.go:117] "RemoveContainer" containerID="50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.110415 4935 scope.go:117] "RemoveContainer" containerID="a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.125781 4935 scope.go:117] "RemoveContainer" containerID="a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52" Oct 05 06:56:24 crc kubenswrapper[4935]: E1005 06:56:24.126260 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52\": container with ID starting with a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52 not found: ID does not exist" containerID="a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.126289 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52"} err="failed to get container status \"a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52\": rpc error: code = NotFound desc = could not find container \"a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52\": container with ID starting with a4c7986203aacbd12782cc559338d64eeec83dc535bb593452ae8bbf580b6e52 not found: ID does not exist" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.126310 4935 scope.go:117] "RemoveContainer" containerID="50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9" Oct 05 06:56:24 crc kubenswrapper[4935]: E1005 06:56:24.126620 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9\": container with ID starting with 50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9 not found: ID does not exist" containerID="50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.126668 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9"} err="failed to get container status \"50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9\": rpc error: code = NotFound desc = could not find container \"50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9\": container with ID starting with 50d6bb11694f75a1d94e33edb00c161a4c27ccdfe14514e257f851f6061237f9 not found: ID does not exist" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.126694 4935 scope.go:117] "RemoveContainer" containerID="a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a" Oct 05 06:56:24 crc kubenswrapper[4935]: E1005 06:56:24.126950 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a\": container with ID starting with a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a not found: ID does not exist" containerID="a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.126975 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a"} err="failed to get container status \"a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a\": rpc error: code = NotFound desc = could not find container \"a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a\": container with ID starting with a3b4336422ed44b91b30e7a71d1791cf4a09d005de4645477bb38b54034f553a not found: ID does not exist" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.127611 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f8beac5-0f55-450a-8d18-d4f331076e26" (UID: "6f8beac5-0f55-450a-8d18-d4f331076e26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.151286 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fjxc9" podStartSLOduration=2.104174107 podStartE2EDuration="1m3.151268166s" podCreationTimestamp="2025-10-05 06:55:21 +0000 UTC" firstStartedPulling="2025-10-05 06:55:22.521014745 +0000 UTC m=+156.403641205" lastFinishedPulling="2025-10-05 06:56:23.568108804 +0000 UTC m=+217.450735264" observedRunningTime="2025-10-05 06:56:24.147835768 +0000 UTC m=+218.030462218" watchObservedRunningTime="2025-10-05 06:56:24.151268166 +0000 UTC m=+218.033894626" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.179806 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24fts\" (UniqueName: \"kubernetes.io/projected/6f8beac5-0f55-450a-8d18-d4f331076e26-kube-api-access-24fts\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.179843 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f8beac5-0f55-450a-8d18-d4f331076e26-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.404768 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7kqbg"] Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.412150 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7kqbg"] Oct 05 06:56:24 crc kubenswrapper[4935]: I1005 06:56:24.788351 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" path="/var/lib/kubelet/pods/6f8beac5-0f55-450a-8d18-d4f331076e26/volumes" Oct 05 06:56:25 crc kubenswrapper[4935]: I1005 06:56:25.091857 4935 generic.go:334] "Generic (PLEG): container finished" podID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerID="8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a" exitCode=0 Oct 05 06:56:25 crc kubenswrapper[4935]: I1005 06:56:25.092034 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rk8k" event={"ID":"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6","Type":"ContainerDied","Data":"8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a"} Oct 05 06:56:25 crc kubenswrapper[4935]: I1005 06:56:25.095082 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8lrt" event={"ID":"05aae77b-1528-4993-9187-a4b4513fcf4f","Type":"ContainerStarted","Data":"0064dd4e9a155ea81ef2c38696b2b597a0f4f5cbcb09ed2856cfbd23926d70da"} Oct 05 06:56:25 crc kubenswrapper[4935]: I1005 06:56:25.096603 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4scg9" event={"ID":"70f5830f-6e20-4066-9419-7916d6a6c0ca","Type":"ContainerStarted","Data":"f8e58eb1e5728d9c3a8304c8c06cd04de7c6d7a80115af9860187864e708b63c"} Oct 05 06:56:25 crc kubenswrapper[4935]: I1005 06:56:25.098303 4935 generic.go:334] "Generic (PLEG): container finished" podID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerID="3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64" exitCode=0 Oct 05 06:56:25 crc kubenswrapper[4935]: I1005 06:56:25.098335 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rmsz" event={"ID":"ce5d3e82-fc79-483f-8e45-8fc712891e7c","Type":"ContainerDied","Data":"3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64"} Oct 05 06:56:25 crc kubenswrapper[4935]: I1005 06:56:25.175294 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v8lrt" podStartSLOduration=3.290141564 podStartE2EDuration="1m4.175279308s" podCreationTimestamp="2025-10-05 06:55:21 +0000 UTC" firstStartedPulling="2025-10-05 06:55:23.591286653 +0000 UTC m=+157.473913113" lastFinishedPulling="2025-10-05 06:56:24.476424387 +0000 UTC m=+218.359050857" observedRunningTime="2025-10-05 06:56:25.172164079 +0000 UTC m=+219.054790539" watchObservedRunningTime="2025-10-05 06:56:25.175279308 +0000 UTC m=+219.057905768" Oct 05 06:56:26 crc kubenswrapper[4935]: I1005 06:56:26.109685 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rk8k" event={"ID":"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6","Type":"ContainerStarted","Data":"3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35"} Oct 05 06:56:26 crc kubenswrapper[4935]: I1005 06:56:26.112981 4935 generic.go:334] "Generic (PLEG): container finished" podID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerID="f8e58eb1e5728d9c3a8304c8c06cd04de7c6d7a80115af9860187864e708b63c" exitCode=0 Oct 05 06:56:26 crc kubenswrapper[4935]: I1005 06:56:26.113063 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4scg9" event={"ID":"70f5830f-6e20-4066-9419-7916d6a6c0ca","Type":"ContainerDied","Data":"f8e58eb1e5728d9c3a8304c8c06cd04de7c6d7a80115af9860187864e708b63c"} Oct 05 06:56:26 crc kubenswrapper[4935]: I1005 06:56:26.118617 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rmsz" event={"ID":"ce5d3e82-fc79-483f-8e45-8fc712891e7c","Type":"ContainerStarted","Data":"6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b"} Oct 05 06:56:26 crc kubenswrapper[4935]: I1005 06:56:26.135561 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6rk8k" podStartSLOduration=2.941447926 podStartE2EDuration="1m7.135538728s" podCreationTimestamp="2025-10-05 06:55:19 +0000 UTC" firstStartedPulling="2025-10-05 06:55:21.456422191 +0000 UTC m=+155.339048651" lastFinishedPulling="2025-10-05 06:56:25.650512983 +0000 UTC m=+219.533139453" observedRunningTime="2025-10-05 06:56:26.131827941 +0000 UTC m=+220.014454431" watchObservedRunningTime="2025-10-05 06:56:26.135538728 +0000 UTC m=+220.018165208" Oct 05 06:56:26 crc kubenswrapper[4935]: I1005 06:56:26.213329 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8rmsz" podStartSLOduration=2.241892879 podStartE2EDuration="1m4.213313543s" podCreationTimestamp="2025-10-05 06:55:22 +0000 UTC" firstStartedPulling="2025-10-05 06:55:23.588866518 +0000 UTC m=+157.471492978" lastFinishedPulling="2025-10-05 06:56:25.560287182 +0000 UTC m=+219.442913642" observedRunningTime="2025-10-05 06:56:26.211257784 +0000 UTC m=+220.093884244" watchObservedRunningTime="2025-10-05 06:56:26.213313543 +0000 UTC m=+220.095940003" Oct 05 06:56:27 crc kubenswrapper[4935]: I1005 06:56:27.126213 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4scg9" event={"ID":"70f5830f-6e20-4066-9419-7916d6a6c0ca","Type":"ContainerStarted","Data":"b97616edf91bded2b576fb2a61115ac4209c44348a7484cd1238ea89aed537e1"} Oct 05 06:56:27 crc kubenswrapper[4935]: I1005 06:56:27.144155 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4scg9" podStartSLOduration=3.200529602 podStartE2EDuration="1m5.1441415s" podCreationTimestamp="2025-10-05 06:55:22 +0000 UTC" firstStartedPulling="2025-10-05 06:55:24.592472158 +0000 UTC m=+158.475098608" lastFinishedPulling="2025-10-05 06:56:26.536084046 +0000 UTC m=+220.418710506" observedRunningTime="2025-10-05 06:56:27.143810491 +0000 UTC m=+221.026436951" watchObservedRunningTime="2025-10-05 06:56:27.1441415 +0000 UTC m=+221.026767960" Oct 05 06:56:29 crc kubenswrapper[4935]: I1005 06:56:29.549091 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:56:29 crc kubenswrapper[4935]: I1005 06:56:29.549588 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:56:29 crc kubenswrapper[4935]: I1005 06:56:29.598587 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:56:29 crc kubenswrapper[4935]: I1005 06:56:29.728222 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:56:29 crc kubenswrapper[4935]: I1005 06:56:29.728267 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:56:29 crc kubenswrapper[4935]: I1005 06:56:29.768868 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:56:29 crc kubenswrapper[4935]: I1005 06:56:29.981080 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:56:29 crc kubenswrapper[4935]: I1005 06:56:29.981145 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:56:30 crc kubenswrapper[4935]: I1005 06:56:30.016281 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:56:30 crc kubenswrapper[4935]: I1005 06:56:30.182816 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:56:30 crc kubenswrapper[4935]: I1005 06:56:30.202232 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:56:30 crc kubenswrapper[4935]: I1005 06:56:30.207994 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:56:31 crc kubenswrapper[4935]: I1005 06:56:31.215955 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9hwvw"] Oct 05 06:56:31 crc kubenswrapper[4935]: I1005 06:56:31.726537 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:56:31 crc kubenswrapper[4935]: I1005 06:56:31.726834 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:56:31 crc kubenswrapper[4935]: I1005 06:56:31.772375 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.005598 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wzkhf"] Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.152052 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wzkhf" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerName="registry-server" containerID="cri-o://02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe" gracePeriod=2 Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.153802 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.153835 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.198883 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.200425 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.526285 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.699148 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-catalog-content\") pod \"30c3e073-685a-4fe1-bdb3-12289e64075b\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.699232 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htqlq\" (UniqueName: \"kubernetes.io/projected/30c3e073-685a-4fe1-bdb3-12289e64075b-kube-api-access-htqlq\") pod \"30c3e073-685a-4fe1-bdb3-12289e64075b\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.699306 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-utilities\") pod \"30c3e073-685a-4fe1-bdb3-12289e64075b\" (UID: \"30c3e073-685a-4fe1-bdb3-12289e64075b\") " Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.700283 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-utilities" (OuterVolumeSpecName: "utilities") pod "30c3e073-685a-4fe1-bdb3-12289e64075b" (UID: "30c3e073-685a-4fe1-bdb3-12289e64075b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.706417 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30c3e073-685a-4fe1-bdb3-12289e64075b-kube-api-access-htqlq" (OuterVolumeSpecName: "kube-api-access-htqlq") pod "30c3e073-685a-4fe1-bdb3-12289e64075b" (UID: "30c3e073-685a-4fe1-bdb3-12289e64075b"). InnerVolumeSpecName "kube-api-access-htqlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.745653 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.745696 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.753025 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30c3e073-685a-4fe1-bdb3-12289e64075b" (UID: "30c3e073-685a-4fe1-bdb3-12289e64075b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.796492 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.800373 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.800396 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htqlq\" (UniqueName: \"kubernetes.io/projected/30c3e073-685a-4fe1-bdb3-12289e64075b-kube-api-access-htqlq\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:32 crc kubenswrapper[4935]: I1005 06:56:32.800408 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c3e073-685a-4fe1-bdb3-12289e64075b-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.159405 4935 generic.go:334] "Generic (PLEG): container finished" podID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerID="02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe" exitCode=0 Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.159514 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzkhf" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.159541 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzkhf" event={"ID":"30c3e073-685a-4fe1-bdb3-12289e64075b","Type":"ContainerDied","Data":"02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe"} Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.159587 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzkhf" event={"ID":"30c3e073-685a-4fe1-bdb3-12289e64075b","Type":"ContainerDied","Data":"f5d74bba98044d41cd81fff7e1f0a2f7b815b3b65841f8fde0d4768dc00921b4"} Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.159608 4935 scope.go:117] "RemoveContainer" containerID="02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.195536 4935 scope.go:117] "RemoveContainer" containerID="d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.200663 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wzkhf"] Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.202933 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.203473 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wzkhf"] Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.217536 4935 scope.go:117] "RemoveContainer" containerID="cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.217736 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.232258 4935 scope.go:117] "RemoveContainer" containerID="02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe" Oct 05 06:56:33 crc kubenswrapper[4935]: E1005 06:56:33.232816 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe\": container with ID starting with 02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe not found: ID does not exist" containerID="02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.232856 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe"} err="failed to get container status \"02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe\": rpc error: code = NotFound desc = could not find container \"02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe\": container with ID starting with 02abfc9c2a0db12d04d0fae985c235ea4945d1e2981bf1b9c33298a85d1673fe not found: ID does not exist" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.232884 4935 scope.go:117] "RemoveContainer" containerID="d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e" Oct 05 06:56:33 crc kubenswrapper[4935]: E1005 06:56:33.233234 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e\": container with ID starting with d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e not found: ID does not exist" containerID="d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.233256 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e"} err="failed to get container status \"d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e\": rpc error: code = NotFound desc = could not find container \"d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e\": container with ID starting with d1c4db65658d8bc8ce46ee3e0e9ec4546e06766f8cf4151eae9971c2e28a715e not found: ID does not exist" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.233274 4935 scope.go:117] "RemoveContainer" containerID="cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077" Oct 05 06:56:33 crc kubenswrapper[4935]: E1005 06:56:33.233577 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077\": container with ID starting with cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077 not found: ID does not exist" containerID="cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.233609 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077"} err="failed to get container status \"cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077\": rpc error: code = NotFound desc = could not find container \"cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077\": container with ID starting with cd0b69bee406a45d1fd036ef44f2781ffb436719b35052039dadbb50db19b077 not found: ID does not exist" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.244073 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.244134 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:56:33 crc kubenswrapper[4935]: I1005 06:56:33.282766 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:56:34 crc kubenswrapper[4935]: I1005 06:56:34.214177 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:56:34 crc kubenswrapper[4935]: I1005 06:56:34.409974 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v8lrt"] Oct 05 06:56:34 crc kubenswrapper[4935]: I1005 06:56:34.798508 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" path="/var/lib/kubelet/pods/30c3e073-685a-4fe1-bdb3-12289e64075b/volumes" Oct 05 06:56:35 crc kubenswrapper[4935]: I1005 06:56:35.171407 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v8lrt" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerName="registry-server" containerID="cri-o://0064dd4e9a155ea81ef2c38696b2b597a0f4f5cbcb09ed2856cfbd23926d70da" gracePeriod=2 Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.178820 4935 generic.go:334] "Generic (PLEG): container finished" podID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerID="0064dd4e9a155ea81ef2c38696b2b597a0f4f5cbcb09ed2856cfbd23926d70da" exitCode=0 Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.178877 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8lrt" event={"ID":"05aae77b-1528-4993-9187-a4b4513fcf4f","Type":"ContainerDied","Data":"0064dd4e9a155ea81ef2c38696b2b597a0f4f5cbcb09ed2856cfbd23926d70da"} Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.715514 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.807692 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4scg9"] Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.807966 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4scg9" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerName="registry-server" containerID="cri-o://b97616edf91bded2b576fb2a61115ac4209c44348a7484cd1238ea89aed537e1" gracePeriod=2 Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.847270 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfzw7\" (UniqueName: \"kubernetes.io/projected/05aae77b-1528-4993-9187-a4b4513fcf4f-kube-api-access-wfzw7\") pod \"05aae77b-1528-4993-9187-a4b4513fcf4f\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.847560 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-utilities\") pod \"05aae77b-1528-4993-9187-a4b4513fcf4f\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.847612 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-catalog-content\") pod \"05aae77b-1528-4993-9187-a4b4513fcf4f\" (UID: \"05aae77b-1528-4993-9187-a4b4513fcf4f\") " Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.848666 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-utilities" (OuterVolumeSpecName: "utilities") pod "05aae77b-1528-4993-9187-a4b4513fcf4f" (UID: "05aae77b-1528-4993-9187-a4b4513fcf4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.852313 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05aae77b-1528-4993-9187-a4b4513fcf4f-kube-api-access-wfzw7" (OuterVolumeSpecName: "kube-api-access-wfzw7") pod "05aae77b-1528-4993-9187-a4b4513fcf4f" (UID: "05aae77b-1528-4993-9187-a4b4513fcf4f"). InnerVolumeSpecName "kube-api-access-wfzw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.864524 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05aae77b-1528-4993-9187-a4b4513fcf4f" (UID: "05aae77b-1528-4993-9187-a4b4513fcf4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.948486 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfzw7\" (UniqueName: \"kubernetes.io/projected/05aae77b-1528-4993-9187-a4b4513fcf4f-kube-api-access-wfzw7\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.948516 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:36 crc kubenswrapper[4935]: I1005 06:56:36.948525 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05aae77b-1528-4993-9187-a4b4513fcf4f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:37 crc kubenswrapper[4935]: I1005 06:56:37.192477 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8lrt" event={"ID":"05aae77b-1528-4993-9187-a4b4513fcf4f","Type":"ContainerDied","Data":"b3a4affb976ac5efd631e0ece66f4e79dfb8292043cb9ec92effd93640b07973"} Oct 05 06:56:37 crc kubenswrapper[4935]: I1005 06:56:37.192548 4935 scope.go:117] "RemoveContainer" containerID="0064dd4e9a155ea81ef2c38696b2b597a0f4f5cbcb09ed2856cfbd23926d70da" Oct 05 06:56:37 crc kubenswrapper[4935]: I1005 06:56:37.192567 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v8lrt" Oct 05 06:56:37 crc kubenswrapper[4935]: I1005 06:56:37.204880 4935 scope.go:117] "RemoveContainer" containerID="cd9b97de52276d24102887bef5e1847e155b70503bfa04296412f02f3e4f1be0" Oct 05 06:56:37 crc kubenswrapper[4935]: I1005 06:56:37.219086 4935 scope.go:117] "RemoveContainer" containerID="f3159b808d14b5719191980922c6a426c7f7a1659da952a3d14c4d31b4f0a43b" Oct 05 06:56:37 crc kubenswrapper[4935]: I1005 06:56:37.261291 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v8lrt"] Oct 05 06:56:37 crc kubenswrapper[4935]: I1005 06:56:37.263712 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v8lrt"] Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.201560 4935 generic.go:334] "Generic (PLEG): container finished" podID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerID="b97616edf91bded2b576fb2a61115ac4209c44348a7484cd1238ea89aed537e1" exitCode=0 Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.201599 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4scg9" event={"ID":"70f5830f-6e20-4066-9419-7916d6a6c0ca","Type":"ContainerDied","Data":"b97616edf91bded2b576fb2a61115ac4209c44348a7484cd1238ea89aed537e1"} Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.583924 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.690204 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-utilities\") pod \"70f5830f-6e20-4066-9419-7916d6a6c0ca\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.690285 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-catalog-content\") pod \"70f5830f-6e20-4066-9419-7916d6a6c0ca\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.690356 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj842\" (UniqueName: \"kubernetes.io/projected/70f5830f-6e20-4066-9419-7916d6a6c0ca-kube-api-access-gj842\") pod \"70f5830f-6e20-4066-9419-7916d6a6c0ca\" (UID: \"70f5830f-6e20-4066-9419-7916d6a6c0ca\") " Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.691318 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-utilities" (OuterVolumeSpecName: "utilities") pod "70f5830f-6e20-4066-9419-7916d6a6c0ca" (UID: "70f5830f-6e20-4066-9419-7916d6a6c0ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.695392 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f5830f-6e20-4066-9419-7916d6a6c0ca-kube-api-access-gj842" (OuterVolumeSpecName: "kube-api-access-gj842") pod "70f5830f-6e20-4066-9419-7916d6a6c0ca" (UID: "70f5830f-6e20-4066-9419-7916d6a6c0ca"). InnerVolumeSpecName "kube-api-access-gj842". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.784275 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" path="/var/lib/kubelet/pods/05aae77b-1528-4993-9187-a4b4513fcf4f/volumes" Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.791653 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj842\" (UniqueName: \"kubernetes.io/projected/70f5830f-6e20-4066-9419-7916d6a6c0ca-kube-api-access-gj842\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.791674 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.864633 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70f5830f-6e20-4066-9419-7916d6a6c0ca" (UID: "70f5830f-6e20-4066-9419-7916d6a6c0ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:38 crc kubenswrapper[4935]: I1005 06:56:38.892538 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f5830f-6e20-4066-9419-7916d6a6c0ca-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:39 crc kubenswrapper[4935]: I1005 06:56:39.208622 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4scg9" event={"ID":"70f5830f-6e20-4066-9419-7916d6a6c0ca","Type":"ContainerDied","Data":"ba08060bc3b3d1d9a83c8084c67a52ab5cbf9476a0ef56cf88db24bfed0e6a47"} Oct 05 06:56:39 crc kubenswrapper[4935]: I1005 06:56:39.208671 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4scg9" Oct 05 06:56:39 crc kubenswrapper[4935]: I1005 06:56:39.208701 4935 scope.go:117] "RemoveContainer" containerID="b97616edf91bded2b576fb2a61115ac4209c44348a7484cd1238ea89aed537e1" Oct 05 06:56:39 crc kubenswrapper[4935]: I1005 06:56:39.238000 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4scg9"] Oct 05 06:56:39 crc kubenswrapper[4935]: I1005 06:56:39.241900 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4scg9"] Oct 05 06:56:39 crc kubenswrapper[4935]: I1005 06:56:39.242946 4935 scope.go:117] "RemoveContainer" containerID="f8e58eb1e5728d9c3a8304c8c06cd04de7c6d7a80115af9860187864e708b63c" Oct 05 06:56:39 crc kubenswrapper[4935]: I1005 06:56:39.260075 4935 scope.go:117] "RemoveContainer" containerID="729342ad817013847f0be059797677d7236d3eee5193aa794056416d1c60ce2e" Oct 05 06:56:40 crc kubenswrapper[4935]: I1005 06:56:40.784146 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" path="/var/lib/kubelet/pods/70f5830f-6e20-4066-9419-7916d6a6c0ca/volumes" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.254375 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" podUID="14973180-053d-4f7b-9f72-7acc0f8b19ac" containerName="oauth-openshift" containerID="cri-o://7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf" gracePeriod=15 Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.772528 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815331 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz"] Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815562 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerName="extract-utilities" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815578 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerName="extract-utilities" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815588 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerName="extract-utilities" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815596 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerName="extract-utilities" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815613 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815621 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815634 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerName="extract-content" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815642 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerName="extract-content" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815652 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba77708b-bb6a-4b80-bfa2-03f36b142595" containerName="pruner" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815660 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba77708b-bb6a-4b80-bfa2-03f36b142595" containerName="pruner" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815670 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerName="extract-content" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815678 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerName="extract-content" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815687 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerName="extract-utilities" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815695 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerName="extract-utilities" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815706 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerName="extract-content" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815714 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerName="extract-content" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815735 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815742 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815752 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815760 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815772 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerName="extract-utilities" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815781 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerName="extract-utilities" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815793 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerName="extract-content" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815801 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerName="extract-content" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815812 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14973180-053d-4f7b-9f72-7acc0f8b19ac" containerName="oauth-openshift" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815820 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="14973180-053d-4f7b-9f72-7acc0f8b19ac" containerName="oauth-openshift" Oct 05 06:56:56 crc kubenswrapper[4935]: E1005 06:56:56.815832 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815840 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815977 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="30c3e073-685a-4fe1-bdb3-12289e64075b" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.815995 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba77708b-bb6a-4b80-bfa2-03f36b142595" containerName="pruner" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.816007 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f8beac5-0f55-450a-8d18-d4f331076e26" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.816017 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="14973180-053d-4f7b-9f72-7acc0f8b19ac" containerName="oauth-openshift" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.816029 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="05aae77b-1528-4993-9187-a4b4513fcf4f" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.816040 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f5830f-6e20-4066-9419-7916d6a6c0ca" containerName="registry-server" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.816492 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.829140 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz"] Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.939343 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.939215 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-dir\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.939583 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-cliconfig\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.941069 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.941237 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-login\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.941328 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-error\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942254 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-session\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942325 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-provider-selection\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942376 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-trusted-ca-bundle\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942421 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-ocp-branding-template\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942622 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-service-ca\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942679 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-router-certs\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942730 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nxrl\" (UniqueName: \"kubernetes.io/projected/14973180-053d-4f7b-9f72-7acc0f8b19ac-kube-api-access-9nxrl\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942806 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-serving-cert\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.942846 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-policies\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943194 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-idp-0-file-data\") pod \"14973180-053d-4f7b-9f72-7acc0f8b19ac\" (UID: \"14973180-053d-4f7b-9f72-7acc0f8b19ac\") " Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943435 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-serving-cert\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943497 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-login\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943545 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943589 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6356d23-3e52-44a6-8aa6-4edd22a54e34-audit-dir\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943626 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-cliconfig\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943699 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943664 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-audit-policies\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943865 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.943977 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-session\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944027 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-router-certs\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944073 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-service-ca\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944152 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944219 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944307 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r22rp\" (UniqueName: \"kubernetes.io/projected/b6356d23-3e52-44a6-8aa6-4edd22a54e34-kube-api-access-r22rp\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944357 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-error\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944620 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944668 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944683 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944859 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.944936 4935 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.949819 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.950434 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14973180-053d-4f7b-9f72-7acc0f8b19ac-kube-api-access-9nxrl" (OuterVolumeSpecName: "kube-api-access-9nxrl") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "kube-api-access-9nxrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.951052 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.952842 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.953529 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.953757 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.954262 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.956090 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:56 crc kubenswrapper[4935]: I1005 06:56:56.957205 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "14973180-053d-4f7b-9f72-7acc0f8b19ac" (UID: "14973180-053d-4f7b-9f72-7acc0f8b19ac"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.045978 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-serving-cert\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046057 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-login\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046089 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046114 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6356d23-3e52-44a6-8aa6-4edd22a54e34-audit-dir\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046201 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-cliconfig\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046225 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-audit-policies\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046252 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046287 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-session\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046313 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-router-certs\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046338 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-service-ca\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046380 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046444 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046470 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r22rp\" (UniqueName: \"kubernetes.io/projected/b6356d23-3e52-44a6-8aa6-4edd22a54e34-kube-api-access-r22rp\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046497 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-error\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046556 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046579 4935 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046591 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046603 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046614 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046625 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046639 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046654 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046674 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046688 4935 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14973180-053d-4f7b-9f72-7acc0f8b19ac-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.046710 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nxrl\" (UniqueName: \"kubernetes.io/projected/14973180-053d-4f7b-9f72-7acc0f8b19ac-kube-api-access-9nxrl\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.050336 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6356d23-3e52-44a6-8aa6-4edd22a54e34-audit-dir\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.050514 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-cliconfig\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.050695 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-audit-policies\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.050921 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-service-ca\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.051911 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-session\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.052434 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-login\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.053474 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-error\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.053757 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.054820 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.055239 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-serving-cert\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.055561 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-router-certs\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.056350 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.058818 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b6356d23-3e52-44a6-8aa6-4edd22a54e34-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.068816 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r22rp\" (UniqueName: \"kubernetes.io/projected/b6356d23-3e52-44a6-8aa6-4edd22a54e34-kube-api-access-r22rp\") pod \"oauth-openshift-679cb4ddc5-tfdsz\" (UID: \"b6356d23-3e52-44a6-8aa6-4edd22a54e34\") " pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.146179 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.326621 4935 generic.go:334] "Generic (PLEG): container finished" podID="14973180-053d-4f7b-9f72-7acc0f8b19ac" containerID="7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf" exitCode=0 Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.326677 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" event={"ID":"14973180-053d-4f7b-9f72-7acc0f8b19ac","Type":"ContainerDied","Data":"7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf"} Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.326717 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" event={"ID":"14973180-053d-4f7b-9f72-7acc0f8b19ac","Type":"ContainerDied","Data":"559dae40682f56604252d745d43ba83d9e5d57e959267f82229aac17941f4d53"} Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.326729 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9hwvw" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.326743 4935 scope.go:117] "RemoveContainer" containerID="7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.357539 4935 scope.go:117] "RemoveContainer" containerID="7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf" Oct 05 06:56:57 crc kubenswrapper[4935]: E1005 06:56:57.358562 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf\": container with ID starting with 7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf not found: ID does not exist" containerID="7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.358615 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf"} err="failed to get container status \"7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf\": rpc error: code = NotFound desc = could not find container \"7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf\": container with ID starting with 7e4d8022e55725d91659804e2aadf6b4e12ec8a9754511a47dd0088d4369a7bf not found: ID does not exist" Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.360781 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9hwvw"] Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.429364 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9hwvw"] Oct 05 06:56:57 crc kubenswrapper[4935]: I1005 06:56:57.451034 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz"] Oct 05 06:56:58 crc kubenswrapper[4935]: I1005 06:56:58.336848 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" event={"ID":"b6356d23-3e52-44a6-8aa6-4edd22a54e34","Type":"ContainerStarted","Data":"ffdc80d11f4d298cfeb79f7e082f3ca7969054937362901b7e9ec446b4f975ff"} Oct 05 06:56:58 crc kubenswrapper[4935]: I1005 06:56:58.336926 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" event={"ID":"b6356d23-3e52-44a6-8aa6-4edd22a54e34","Type":"ContainerStarted","Data":"78c3cd6a5d24378febfd8484eab042456bf6e33e8c47294b119435a5fb15c94d"} Oct 05 06:56:58 crc kubenswrapper[4935]: I1005 06:56:58.337591 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:58 crc kubenswrapper[4935]: I1005 06:56:58.348527 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" Oct 05 06:56:58 crc kubenswrapper[4935]: I1005 06:56:58.378782 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-679cb4ddc5-tfdsz" podStartSLOduration=27.378755512 podStartE2EDuration="27.378755512s" podCreationTimestamp="2025-10-05 06:56:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:56:58.37066408 +0000 UTC m=+252.253290560" watchObservedRunningTime="2025-10-05 06:56:58.378755512 +0000 UTC m=+252.261381992" Oct 05 06:56:58 crc kubenswrapper[4935]: I1005 06:56:58.788196 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14973180-053d-4f7b-9f72-7acc0f8b19ac" path="/var/lib/kubelet/pods/14973180-053d-4f7b-9f72-7acc0f8b19ac/volumes" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.495819 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6rk8k"] Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.496747 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6rk8k" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerName="registry-server" containerID="cri-o://3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35" gracePeriod=30 Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.506948 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4q5n"] Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.510908 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvn45"] Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.513218 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" podUID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" containerName="marketplace-operator" containerID="cri-o://e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9" gracePeriod=30 Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.525570 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjxc9"] Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.525786 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fjxc9" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerName="registry-server" containerID="cri-o://9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce" gracePeriod=30 Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.536361 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8rmsz"] Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.536579 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8rmsz" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerName="registry-server" containerID="cri-o://6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b" gracePeriod=30 Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.543800 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g9n7m"] Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.546801 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.573867 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g9n7m"] Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.614272 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6067d7a8-863b-46f1-97d0-e848aea36ebf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.614318 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6067d7a8-863b-46f1-97d0-e848aea36ebf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.614341 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp7pg\" (UniqueName: \"kubernetes.io/projected/6067d7a8-863b-46f1-97d0-e848aea36ebf-kube-api-access-sp7pg\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.617904 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p4q5n" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerName="registry-server" containerID="cri-o://9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252" gracePeriod=30 Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.716240 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp7pg\" (UniqueName: \"kubernetes.io/projected/6067d7a8-863b-46f1-97d0-e848aea36ebf-kube-api-access-sp7pg\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.716343 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6067d7a8-863b-46f1-97d0-e848aea36ebf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.716387 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6067d7a8-863b-46f1-97d0-e848aea36ebf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.719066 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6067d7a8-863b-46f1-97d0-e848aea36ebf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.723784 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6067d7a8-863b-46f1-97d0-e848aea36ebf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.733257 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp7pg\" (UniqueName: \"kubernetes.io/projected/6067d7a8-863b-46f1-97d0-e848aea36ebf-kube-api-access-sp7pg\") pod \"marketplace-operator-79b997595-g9n7m\" (UID: \"6067d7a8-863b-46f1-97d0-e848aea36ebf\") " pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.922488 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:34 crc kubenswrapper[4935]: I1005 06:57:34.927472 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.018046 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.026773 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-trusted-ca\") pod \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.026829 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s9dn\" (UniqueName: \"kubernetes.io/projected/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-kube-api-access-9s9dn\") pod \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.026875 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-operator-metrics\") pod \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\" (UID: \"95dcac01-54e5-4fcd-a4eb-068edbe7da4f\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.028154 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "95dcac01-54e5-4fcd-a4eb-068edbe7da4f" (UID: "95dcac01-54e5-4fcd-a4eb-068edbe7da4f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.030693 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-kube-api-access-9s9dn" (OuterVolumeSpecName: "kube-api-access-9s9dn") pod "95dcac01-54e5-4fcd-a4eb-068edbe7da4f" (UID: "95dcac01-54e5-4fcd-a4eb-068edbe7da4f"). InnerVolumeSpecName "kube-api-access-9s9dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.031968 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.035231 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "95dcac01-54e5-4fcd-a4eb-068edbe7da4f" (UID: "95dcac01-54e5-4fcd-a4eb-068edbe7da4f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.038819 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.057242 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.127986 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-utilities\") pod \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128025 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-utilities\") pod \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128074 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqkmx\" (UniqueName: \"kubernetes.io/projected/a1476c09-cca4-4cd9-adfd-188e00dc9457-kube-api-access-tqkmx\") pod \"a1476c09-cca4-4cd9-adfd-188e00dc9457\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128093 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxkwc\" (UniqueName: \"kubernetes.io/projected/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-kube-api-access-zxkwc\") pod \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128113 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-catalog-content\") pod \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\" (UID: \"638ea9d6-e6c3-4ae4-b361-31ca73cb242a\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128135 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-utilities\") pod \"a1476c09-cca4-4cd9-adfd-188e00dc9457\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128168 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-catalog-content\") pod \"a1476c09-cca4-4cd9-adfd-188e00dc9457\" (UID: \"a1476c09-cca4-4cd9-adfd-188e00dc9457\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128190 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-catalog-content\") pod \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128230 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz62k\" (UniqueName: \"kubernetes.io/projected/ce5d3e82-fc79-483f-8e45-8fc712891e7c-kube-api-access-vz62k\") pod \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128250 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v4g9\" (UniqueName: \"kubernetes.io/projected/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-kube-api-access-9v4g9\") pod \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128270 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-utilities\") pod \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\" (UID: \"ce5d3e82-fc79-483f-8e45-8fc712891e7c\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128302 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-catalog-content\") pod \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\" (UID: \"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6\") " Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128487 4935 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128501 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s9dn\" (UniqueName: \"kubernetes.io/projected/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-kube-api-access-9s9dn\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.128512 4935 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95dcac01-54e5-4fcd-a4eb-068edbe7da4f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.130025 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-utilities" (OuterVolumeSpecName: "utilities") pod "638ea9d6-e6c3-4ae4-b361-31ca73cb242a" (UID: "638ea9d6-e6c3-4ae4-b361-31ca73cb242a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.130153 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-utilities" (OuterVolumeSpecName: "utilities") pod "ce5d3e82-fc79-483f-8e45-8fc712891e7c" (UID: "ce5d3e82-fc79-483f-8e45-8fc712891e7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.130753 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-utilities" (OuterVolumeSpecName: "utilities") pod "5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" (UID: "5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.132094 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-utilities" (OuterVolumeSpecName: "utilities") pod "a1476c09-cca4-4cd9-adfd-188e00dc9457" (UID: "a1476c09-cca4-4cd9-adfd-188e00dc9457"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.134008 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5d3e82-fc79-483f-8e45-8fc712891e7c-kube-api-access-vz62k" (OuterVolumeSpecName: "kube-api-access-vz62k") pod "ce5d3e82-fc79-483f-8e45-8fc712891e7c" (UID: "ce5d3e82-fc79-483f-8e45-8fc712891e7c"). InnerVolumeSpecName "kube-api-access-vz62k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.134237 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-kube-api-access-zxkwc" (OuterVolumeSpecName: "kube-api-access-zxkwc") pod "638ea9d6-e6c3-4ae4-b361-31ca73cb242a" (UID: "638ea9d6-e6c3-4ae4-b361-31ca73cb242a"). InnerVolumeSpecName "kube-api-access-zxkwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.137779 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-kube-api-access-9v4g9" (OuterVolumeSpecName: "kube-api-access-9v4g9") pod "5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" (UID: "5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6"). InnerVolumeSpecName "kube-api-access-9v4g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.142275 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "638ea9d6-e6c3-4ae4-b361-31ca73cb242a" (UID: "638ea9d6-e6c3-4ae4-b361-31ca73cb242a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.144717 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1476c09-cca4-4cd9-adfd-188e00dc9457-kube-api-access-tqkmx" (OuterVolumeSpecName: "kube-api-access-tqkmx") pod "a1476c09-cca4-4cd9-adfd-188e00dc9457" (UID: "a1476c09-cca4-4cd9-adfd-188e00dc9457"). InnerVolumeSpecName "kube-api-access-tqkmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.180259 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" (UID: "5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.192150 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1476c09-cca4-4cd9-adfd-188e00dc9457" (UID: "a1476c09-cca4-4cd9-adfd-188e00dc9457"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229381 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v4g9\" (UniqueName: \"kubernetes.io/projected/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-kube-api-access-9v4g9\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229419 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229433 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229444 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229458 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229470 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqkmx\" (UniqueName: \"kubernetes.io/projected/a1476c09-cca4-4cd9-adfd-188e00dc9457-kube-api-access-tqkmx\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229482 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxkwc\" (UniqueName: \"kubernetes.io/projected/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-kube-api-access-zxkwc\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229493 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638ea9d6-e6c3-4ae4-b361-31ca73cb242a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229506 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229517 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1476c09-cca4-4cd9-adfd-188e00dc9457-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.229527 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz62k\" (UniqueName: \"kubernetes.io/projected/ce5d3e82-fc79-483f-8e45-8fc712891e7c-kube-api-access-vz62k\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.240566 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce5d3e82-fc79-483f-8e45-8fc712891e7c" (UID: "ce5d3e82-fc79-483f-8e45-8fc712891e7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.331525 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5d3e82-fc79-483f-8e45-8fc712891e7c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.381535 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g9n7m"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.626563 4935 generic.go:334] "Generic (PLEG): container finished" podID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerID="6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b" exitCode=0 Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.626854 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8rmsz" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.626745 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rmsz" event={"ID":"ce5d3e82-fc79-483f-8e45-8fc712891e7c","Type":"ContainerDied","Data":"6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.626922 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8rmsz" event={"ID":"ce5d3e82-fc79-483f-8e45-8fc712891e7c","Type":"ContainerDied","Data":"808b4260bdd958f059421fd47ea50d6b82b5548dac188c2073c31ddfa3befbd1"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.626951 4935 scope.go:117] "RemoveContainer" containerID="6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.630360 4935 generic.go:334] "Generic (PLEG): container finished" podID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerID="9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252" exitCode=0 Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.630421 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q5n" event={"ID":"a1476c09-cca4-4cd9-adfd-188e00dc9457","Type":"ContainerDied","Data":"9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.630444 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q5n" event={"ID":"a1476c09-cca4-4cd9-adfd-188e00dc9457","Type":"ContainerDied","Data":"c1fd3b37fbc002f3729303639b41030c5849544ee4974563c17d059b6e2d9c27"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.630519 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4q5n" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.647446 4935 scope.go:117] "RemoveContainer" containerID="3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.653177 4935 generic.go:334] "Generic (PLEG): container finished" podID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerID="9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce" exitCode=0 Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.653279 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjxc9" event={"ID":"638ea9d6-e6c3-4ae4-b361-31ca73cb242a","Type":"ContainerDied","Data":"9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.653316 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjxc9" event={"ID":"638ea9d6-e6c3-4ae4-b361-31ca73cb242a","Type":"ContainerDied","Data":"fa83e0c92b23205de8c2257539a8baac2785cb1d666d81c8e365c9a0dc9de35a"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.653414 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjxc9" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.657642 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" event={"ID":"6067d7a8-863b-46f1-97d0-e848aea36ebf","Type":"ContainerStarted","Data":"39a520e7eb97c94ac2b822de4a9735b8737ff5d8915a142ac8f00d94fab3679c"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.657698 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" event={"ID":"6067d7a8-863b-46f1-97d0-e848aea36ebf","Type":"ContainerStarted","Data":"57acae65f15642044245f65e4dccc64992916d8709d592d3a3238e4d6cbb1927"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.658605 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.662573 4935 generic.go:334] "Generic (PLEG): container finished" podID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" containerID="e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9" exitCode=0 Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.662695 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.663027 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8rmsz"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.663139 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" event={"ID":"95dcac01-54e5-4fcd-a4eb-068edbe7da4f","Type":"ContainerDied","Data":"e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.663290 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cvn45" event={"ID":"95dcac01-54e5-4fcd-a4eb-068edbe7da4f","Type":"ContainerDied","Data":"c63501f154fdd7d92bb230d51540306d09be2404bde9c5e21436bffc000c865c"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.663326 4935 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-g9n7m container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.663379 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" podUID="6067d7a8-863b-46f1-97d0-e848aea36ebf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.665590 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8rmsz"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.669943 4935 generic.go:334] "Generic (PLEG): container finished" podID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerID="3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35" exitCode=0 Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.669989 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rk8k" event={"ID":"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6","Type":"ContainerDied","Data":"3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.670022 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6rk8k" event={"ID":"5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6","Type":"ContainerDied","Data":"d2238137491d4f91f2ab0a759e3f49cba805a9bef9d67a461d504976b6a5898c"} Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.670094 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6rk8k" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.688877 4935 scope.go:117] "RemoveContainer" containerID="2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.690038 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" podStartSLOduration=1.690018228 podStartE2EDuration="1.690018228s" podCreationTimestamp="2025-10-05 06:57:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:57:35.686471477 +0000 UTC m=+289.569097957" watchObservedRunningTime="2025-10-05 06:57:35.690018228 +0000 UTC m=+289.572644698" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.708016 4935 scope.go:117] "RemoveContainer" containerID="6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.708511 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b\": container with ID starting with 6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b not found: ID does not exist" containerID="6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.708564 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b"} err="failed to get container status \"6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b\": rpc error: code = NotFound desc = could not find container \"6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b\": container with ID starting with 6aef2ca8c2bbca445fbc2a2d9b977ae125411824b8dc8277b84818f0c1cbfe3b not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.708599 4935 scope.go:117] "RemoveContainer" containerID="3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.709001 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64\": container with ID starting with 3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64 not found: ID does not exist" containerID="3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.709037 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64"} err="failed to get container status \"3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64\": rpc error: code = NotFound desc = could not find container \"3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64\": container with ID starting with 3d03038b5e5e15bf6e73f5c6c2294bebe3f47cf350f38a98456dd9f91b8cfb64 not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.709063 4935 scope.go:117] "RemoveContainer" containerID="2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.709362 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748\": container with ID starting with 2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748 not found: ID does not exist" containerID="2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.709382 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748"} err="failed to get container status \"2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748\": rpc error: code = NotFound desc = could not find container \"2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748\": container with ID starting with 2a32120cdd4bdebc7430330bf6eab7583c31457f5eb04c2807b7cdaacbb25748 not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.709394 4935 scope.go:117] "RemoveContainer" containerID="9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.731170 4935 scope.go:117] "RemoveContainer" containerID="06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.742712 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4q5n"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.747374 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p4q5n"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.754087 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjxc9"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.758676 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjxc9"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.764382 4935 scope.go:117] "RemoveContainer" containerID="5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.769508 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvn45"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.772461 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cvn45"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.775866 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6rk8k"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.778551 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6rk8k"] Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.786009 4935 scope.go:117] "RemoveContainer" containerID="9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.786529 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252\": container with ID starting with 9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252 not found: ID does not exist" containerID="9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.786574 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252"} err="failed to get container status \"9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252\": rpc error: code = NotFound desc = could not find container \"9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252\": container with ID starting with 9e1771ade45511719de0b12e2a3876130bbccdbd23db9905952643706e769252 not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.786607 4935 scope.go:117] "RemoveContainer" containerID="06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.787007 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6\": container with ID starting with 06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6 not found: ID does not exist" containerID="06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.787028 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6"} err="failed to get container status \"06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6\": rpc error: code = NotFound desc = could not find container \"06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6\": container with ID starting with 06c40026214503df198603082cd66191dd5467dcfdbdae8821057ad12eeab0e6 not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.787041 4935 scope.go:117] "RemoveContainer" containerID="5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.787283 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d\": container with ID starting with 5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d not found: ID does not exist" containerID="5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.787312 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d"} err="failed to get container status \"5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d\": rpc error: code = NotFound desc = could not find container \"5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d\": container with ID starting with 5d2bef70c64bdd8fd7426736b6a6f6dc7626c3ce358d5c3ec3ec24842224bd9d not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.787335 4935 scope.go:117] "RemoveContainer" containerID="9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.802430 4935 scope.go:117] "RemoveContainer" containerID="231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.821959 4935 scope.go:117] "RemoveContainer" containerID="a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.836018 4935 scope.go:117] "RemoveContainer" containerID="9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.836481 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce\": container with ID starting with 9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce not found: ID does not exist" containerID="9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.836524 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce"} err="failed to get container status \"9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce\": rpc error: code = NotFound desc = could not find container \"9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce\": container with ID starting with 9840de83cd778fd1cc7d45e819943da4f8c423e5764790407d9cf35c74e18dce not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.836553 4935 scope.go:117] "RemoveContainer" containerID="231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.836851 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7\": container with ID starting with 231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7 not found: ID does not exist" containerID="231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.836882 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7"} err="failed to get container status \"231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7\": rpc error: code = NotFound desc = could not find container \"231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7\": container with ID starting with 231397bbadb5f7c9e13c6e59b845f3bad6608664c073b62f532ea40476025dc7 not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.836927 4935 scope.go:117] "RemoveContainer" containerID="a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.837339 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8\": container with ID starting with a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8 not found: ID does not exist" containerID="a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.837373 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8"} err="failed to get container status \"a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8\": rpc error: code = NotFound desc = could not find container \"a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8\": container with ID starting with a9117fc73356c6bc0a8fd7aa09728b46dc186bc8c878e3453a84bf18f62278d8 not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.837393 4935 scope.go:117] "RemoveContainer" containerID="e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.850153 4935 scope.go:117] "RemoveContainer" containerID="e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.850604 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9\": container with ID starting with e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9 not found: ID does not exist" containerID="e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.850660 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9"} err="failed to get container status \"e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9\": rpc error: code = NotFound desc = could not find container \"e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9\": container with ID starting with e947f856d7610ee262a8a549f114a3dd50baf74a589d6aa1f2bb1d65aa205cc9 not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.850696 4935 scope.go:117] "RemoveContainer" containerID="3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.866074 4935 scope.go:117] "RemoveContainer" containerID="8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.880599 4935 scope.go:117] "RemoveContainer" containerID="b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.892786 4935 scope.go:117] "RemoveContainer" containerID="3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.893459 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35\": container with ID starting with 3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35 not found: ID does not exist" containerID="3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.893512 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35"} err="failed to get container status \"3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35\": rpc error: code = NotFound desc = could not find container \"3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35\": container with ID starting with 3645712ae0b9df61e88853be3f8bfad83bb0ac8139d8f9d9dd7277c336394e35 not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.893549 4935 scope.go:117] "RemoveContainer" containerID="8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.894039 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a\": container with ID starting with 8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a not found: ID does not exist" containerID="8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.894084 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a"} err="failed to get container status \"8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a\": rpc error: code = NotFound desc = could not find container \"8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a\": container with ID starting with 8fd34f28325b211a6f754c26555b4469f392a00fd72ed2aa401c1805f8a3ba8a not found: ID does not exist" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.894114 4935 scope.go:117] "RemoveContainer" containerID="b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac" Oct 05 06:57:35 crc kubenswrapper[4935]: E1005 06:57:35.894458 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac\": container with ID starting with b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac not found: ID does not exist" containerID="b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac" Oct 05 06:57:35 crc kubenswrapper[4935]: I1005 06:57:35.894492 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac"} err="failed to get container status \"b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac\": rpc error: code = NotFound desc = could not find container \"b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac\": container with ID starting with b58e58a6afec0f97fd7404debb9074c143ff6d60441ce1563519c0839c7c36ac not found: ID does not exist" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.684177 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-g9n7m" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700242 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dqf5x"] Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700724 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700743 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700760 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerName="extract-content" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700769 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerName="extract-content" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700783 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerName="extract-utilities" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700794 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerName="extract-utilities" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700808 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerName="extract-utilities" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700817 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerName="extract-utilities" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700830 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerName="extract-utilities" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700839 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerName="extract-utilities" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700850 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerName="extract-content" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700858 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerName="extract-content" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700870 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerName="extract-content" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700879 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerName="extract-content" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700911 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerName="extract-utilities" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700919 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerName="extract-utilities" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700928 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700936 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700950 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerName="extract-content" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700959 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerName="extract-content" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700970 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" containerName="marketplace-operator" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.700978 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" containerName="marketplace-operator" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.700992 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.701000 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: E1005 06:57:36.701010 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.701019 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.701126 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.701147 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.701160 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.701170 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" containerName="marketplace-operator" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.701181 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" containerName="registry-server" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.702064 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.705351 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.713314 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dqf5x"] Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.783145 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6" path="/var/lib/kubelet/pods/5c100f60-a5af-48ef-b2a2-dd0fe04c3ab6/volumes" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.783763 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="638ea9d6-e6c3-4ae4-b361-31ca73cb242a" path="/var/lib/kubelet/pods/638ea9d6-e6c3-4ae4-b361-31ca73cb242a/volumes" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.784445 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95dcac01-54e5-4fcd-a4eb-068edbe7da4f" path="/var/lib/kubelet/pods/95dcac01-54e5-4fcd-a4eb-068edbe7da4f/volumes" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.785467 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1476c09-cca4-4cd9-adfd-188e00dc9457" path="/var/lib/kubelet/pods/a1476c09-cca4-4cd9-adfd-188e00dc9457/volumes" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.786054 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce5d3e82-fc79-483f-8e45-8fc712891e7c" path="/var/lib/kubelet/pods/ce5d3e82-fc79-483f-8e45-8fc712891e7c/volumes" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.853798 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70bc22d6-6824-4a71-ab0b-b739369accbc-catalog-content\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.853849 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4lnc\" (UniqueName: \"kubernetes.io/projected/70bc22d6-6824-4a71-ab0b-b739369accbc-kube-api-access-l4lnc\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.853885 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70bc22d6-6824-4a71-ab0b-b739369accbc-utilities\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.897790 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cl4vw"] Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.900600 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.904864 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.908834 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cl4vw"] Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.954592 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70bc22d6-6824-4a71-ab0b-b739369accbc-catalog-content\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.954629 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4lnc\" (UniqueName: \"kubernetes.io/projected/70bc22d6-6824-4a71-ab0b-b739369accbc-kube-api-access-l4lnc\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.954662 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70bc22d6-6824-4a71-ab0b-b739369accbc-utilities\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.955058 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70bc22d6-6824-4a71-ab0b-b739369accbc-catalog-content\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.955090 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70bc22d6-6824-4a71-ab0b-b739369accbc-utilities\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:36 crc kubenswrapper[4935]: I1005 06:57:36.973825 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4lnc\" (UniqueName: \"kubernetes.io/projected/70bc22d6-6824-4a71-ab0b-b739369accbc-kube-api-access-l4lnc\") pod \"certified-operators-dqf5x\" (UID: \"70bc22d6-6824-4a71-ab0b-b739369accbc\") " pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.026188 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.055857 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zh4p\" (UniqueName: \"kubernetes.io/projected/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-kube-api-access-4zh4p\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.055914 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-utilities\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.055930 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-catalog-content\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.157069 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zh4p\" (UniqueName: \"kubernetes.io/projected/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-kube-api-access-4zh4p\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.157413 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-utilities\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.157437 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-catalog-content\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.157943 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-utilities\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.157996 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-catalog-content\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.176287 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zh4p\" (UniqueName: \"kubernetes.io/projected/9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9-kube-api-access-4zh4p\") pod \"redhat-marketplace-cl4vw\" (UID: \"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9\") " pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.209540 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dqf5x"] Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.226669 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.600511 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cl4vw"] Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.688611 4935 generic.go:334] "Generic (PLEG): container finished" podID="70bc22d6-6824-4a71-ab0b-b739369accbc" containerID="663eb26f594dd206cbd94e38bb094f155e2b7834fa0bcf840ab57f2ccc885c95" exitCode=0 Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.688655 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqf5x" event={"ID":"70bc22d6-6824-4a71-ab0b-b739369accbc","Type":"ContainerDied","Data":"663eb26f594dd206cbd94e38bb094f155e2b7834fa0bcf840ab57f2ccc885c95"} Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.688711 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqf5x" event={"ID":"70bc22d6-6824-4a71-ab0b-b739369accbc","Type":"ContainerStarted","Data":"3f878020695d1a1a5935718717d04ed0d0a1894fb4c67f707e6582a0c2c0ca7c"} Oct 05 06:57:37 crc kubenswrapper[4935]: I1005 06:57:37.691417 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl4vw" event={"ID":"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9","Type":"ContainerStarted","Data":"ae9e87d945bef64a71406fa951f3c8f8c99143ea228e8d7461068f7558e9ebe6"} Oct 05 06:57:38 crc kubenswrapper[4935]: I1005 06:57:38.701276 4935 generic.go:334] "Generic (PLEG): container finished" podID="70bc22d6-6824-4a71-ab0b-b739369accbc" containerID="5bf5f45326fefd1a1a7094866e7f1374b43c6bccd4c96e09f422ada8ead5a757" exitCode=0 Oct 05 06:57:38 crc kubenswrapper[4935]: I1005 06:57:38.701357 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqf5x" event={"ID":"70bc22d6-6824-4a71-ab0b-b739369accbc","Type":"ContainerDied","Data":"5bf5f45326fefd1a1a7094866e7f1374b43c6bccd4c96e09f422ada8ead5a757"} Oct 05 06:57:38 crc kubenswrapper[4935]: I1005 06:57:38.703219 4935 generic.go:334] "Generic (PLEG): container finished" podID="9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9" containerID="be118d59256ea9fc1ff43e45679f1e581612af3cecfc88b7824b177fcb019a2b" exitCode=0 Oct 05 06:57:38 crc kubenswrapper[4935]: I1005 06:57:38.703389 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl4vw" event={"ID":"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9","Type":"ContainerDied","Data":"be118d59256ea9fc1ff43e45679f1e581612af3cecfc88b7824b177fcb019a2b"} Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.101117 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gbnlj"] Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.102803 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.106267 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.113520 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gbnlj"] Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.181365 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btnmq\" (UniqueName: \"kubernetes.io/projected/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-kube-api-access-btnmq\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.181411 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-utilities\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.181440 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-catalog-content\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.282540 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btnmq\" (UniqueName: \"kubernetes.io/projected/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-kube-api-access-btnmq\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.282612 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-utilities\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.282658 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-catalog-content\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.283592 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-catalog-content\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.283704 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-utilities\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.305432 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s65n7"] Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.309143 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.313585 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.315030 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btnmq\" (UniqueName: \"kubernetes.io/projected/d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742-kube-api-access-btnmq\") pod \"redhat-operators-gbnlj\" (UID: \"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742\") " pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.317146 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s65n7"] Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.383735 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljnq2\" (UniqueName: \"kubernetes.io/projected/8007248d-a0fa-4469-97a0-40a224915852-kube-api-access-ljnq2\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.383807 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8007248d-a0fa-4469-97a0-40a224915852-utilities\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.383850 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8007248d-a0fa-4469-97a0-40a224915852-catalog-content\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.427692 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.485418 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8007248d-a0fa-4469-97a0-40a224915852-utilities\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.485494 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8007248d-a0fa-4469-97a0-40a224915852-catalog-content\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.485520 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljnq2\" (UniqueName: \"kubernetes.io/projected/8007248d-a0fa-4469-97a0-40a224915852-kube-api-access-ljnq2\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.486371 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8007248d-a0fa-4469-97a0-40a224915852-utilities\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.486586 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8007248d-a0fa-4469-97a0-40a224915852-catalog-content\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.508872 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljnq2\" (UniqueName: \"kubernetes.io/projected/8007248d-a0fa-4469-97a0-40a224915852-kube-api-access-ljnq2\") pod \"community-operators-s65n7\" (UID: \"8007248d-a0fa-4469-97a0-40a224915852\") " pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.610644 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gbnlj"] Oct 05 06:57:39 crc kubenswrapper[4935]: W1005 06:57:39.613949 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3fc5fe9_197b_45d5_aedd_a1f5bc9e8742.slice/crio-4b5dab037c289d8d6f4ea3ea53424154110110453b8f541d1bb1643977e92127 WatchSource:0}: Error finding container 4b5dab037c289d8d6f4ea3ea53424154110110453b8f541d1bb1643977e92127: Status 404 returned error can't find the container with id 4b5dab037c289d8d6f4ea3ea53424154110110453b8f541d1bb1643977e92127 Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.647926 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.712986 4935 generic.go:334] "Generic (PLEG): container finished" podID="9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9" containerID="5ba43706ecde79afb8e493851f0c9d5833849ff1437d7739105fd8eed4a3b2f1" exitCode=0 Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.713101 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl4vw" event={"ID":"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9","Type":"ContainerDied","Data":"5ba43706ecde79afb8e493851f0c9d5833849ff1437d7739105fd8eed4a3b2f1"} Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.719631 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqf5x" event={"ID":"70bc22d6-6824-4a71-ab0b-b739369accbc","Type":"ContainerStarted","Data":"41ffb6d48004fcb6c682b941a02b83cfd7b6351453379c086ad4c1b52644c1a8"} Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.722383 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbnlj" event={"ID":"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742","Type":"ContainerStarted","Data":"4b5dab037c289d8d6f4ea3ea53424154110110453b8f541d1bb1643977e92127"} Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.751081 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dqf5x" podStartSLOduration=2.310219034 podStartE2EDuration="3.751055993s" podCreationTimestamp="2025-10-05 06:57:36 +0000 UTC" firstStartedPulling="2025-10-05 06:57:37.691015148 +0000 UTC m=+291.573641598" lastFinishedPulling="2025-10-05 06:57:39.131852097 +0000 UTC m=+293.014478557" observedRunningTime="2025-10-05 06:57:39.74841953 +0000 UTC m=+293.631046020" watchObservedRunningTime="2025-10-05 06:57:39.751055993 +0000 UTC m=+293.633682463" Oct 05 06:57:39 crc kubenswrapper[4935]: I1005 06:57:39.862617 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s65n7"] Oct 05 06:57:39 crc kubenswrapper[4935]: W1005 06:57:39.864332 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8007248d_a0fa_4469_97a0_40a224915852.slice/crio-fb87f0bc5da23947f09506077000091bf0c6755a4943791cba545d46473db2de WatchSource:0}: Error finding container fb87f0bc5da23947f09506077000091bf0c6755a4943791cba545d46473db2de: Status 404 returned error can't find the container with id fb87f0bc5da23947f09506077000091bf0c6755a4943791cba545d46473db2de Oct 05 06:57:40 crc kubenswrapper[4935]: I1005 06:57:40.732117 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cl4vw" event={"ID":"9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9","Type":"ContainerStarted","Data":"91f96c7aa82822e9814539e6c8ef5536f2cd073b62eda22ed29a83155b7d5718"} Oct 05 06:57:40 crc kubenswrapper[4935]: I1005 06:57:40.734215 4935 generic.go:334] "Generic (PLEG): container finished" podID="d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742" containerID="36e3d848543c51b73675d71130c3562c49d163d75db11570d655d4f2d281c7d4" exitCode=0 Oct 05 06:57:40 crc kubenswrapper[4935]: I1005 06:57:40.734314 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbnlj" event={"ID":"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742","Type":"ContainerDied","Data":"36e3d848543c51b73675d71130c3562c49d163d75db11570d655d4f2d281c7d4"} Oct 05 06:57:40 crc kubenswrapper[4935]: I1005 06:57:40.736566 4935 generic.go:334] "Generic (PLEG): container finished" podID="8007248d-a0fa-4469-97a0-40a224915852" containerID="db229ebc1c55176f052ee7170427630fccd93b66f629035a20f5cc29fe5edd7b" exitCode=0 Oct 05 06:57:40 crc kubenswrapper[4935]: I1005 06:57:40.737366 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s65n7" event={"ID":"8007248d-a0fa-4469-97a0-40a224915852","Type":"ContainerDied","Data":"db229ebc1c55176f052ee7170427630fccd93b66f629035a20f5cc29fe5edd7b"} Oct 05 06:57:40 crc kubenswrapper[4935]: I1005 06:57:40.737390 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s65n7" event={"ID":"8007248d-a0fa-4469-97a0-40a224915852","Type":"ContainerStarted","Data":"fb87f0bc5da23947f09506077000091bf0c6755a4943791cba545d46473db2de"} Oct 05 06:57:40 crc kubenswrapper[4935]: I1005 06:57:40.752725 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cl4vw" podStartSLOduration=3.08000431 podStartE2EDuration="4.752702739s" podCreationTimestamp="2025-10-05 06:57:36 +0000 UTC" firstStartedPulling="2025-10-05 06:57:38.704737444 +0000 UTC m=+292.587363904" lastFinishedPulling="2025-10-05 06:57:40.377435843 +0000 UTC m=+294.260062333" observedRunningTime="2025-10-05 06:57:40.74988644 +0000 UTC m=+294.632512920" watchObservedRunningTime="2025-10-05 06:57:40.752702739 +0000 UTC m=+294.635329199" Oct 05 06:57:41 crc kubenswrapper[4935]: I1005 06:57:41.748830 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbnlj" event={"ID":"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742","Type":"ContainerStarted","Data":"8cd9da406ac11134276e6402c454b8418d836ca28efdfd3790f971abdc768026"} Oct 05 06:57:41 crc kubenswrapper[4935]: I1005 06:57:41.750934 4935 generic.go:334] "Generic (PLEG): container finished" podID="8007248d-a0fa-4469-97a0-40a224915852" containerID="90ede12533da3f6060f5a4dd09b3500cc46711e16b60f135f00295375f3dcb43" exitCode=0 Oct 05 06:57:41 crc kubenswrapper[4935]: I1005 06:57:41.751008 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s65n7" event={"ID":"8007248d-a0fa-4469-97a0-40a224915852","Type":"ContainerDied","Data":"90ede12533da3f6060f5a4dd09b3500cc46711e16b60f135f00295375f3dcb43"} Oct 05 06:57:42 crc kubenswrapper[4935]: I1005 06:57:42.759436 4935 generic.go:334] "Generic (PLEG): container finished" podID="d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742" containerID="8cd9da406ac11134276e6402c454b8418d836ca28efdfd3790f971abdc768026" exitCode=0 Oct 05 06:57:42 crc kubenswrapper[4935]: I1005 06:57:42.759526 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbnlj" event={"ID":"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742","Type":"ContainerDied","Data":"8cd9da406ac11134276e6402c454b8418d836ca28efdfd3790f971abdc768026"} Oct 05 06:57:42 crc kubenswrapper[4935]: I1005 06:57:42.762244 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s65n7" event={"ID":"8007248d-a0fa-4469-97a0-40a224915852","Type":"ContainerStarted","Data":"f5eb4c2e6154101c7d048b488be83f9eb20f5a73224078a0ca9f67ebc81b03eb"} Oct 05 06:57:42 crc kubenswrapper[4935]: I1005 06:57:42.808365 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s65n7" podStartSLOduration=2.381068123 podStartE2EDuration="3.808343596s" podCreationTimestamp="2025-10-05 06:57:39 +0000 UTC" firstStartedPulling="2025-10-05 06:57:40.740352431 +0000 UTC m=+294.622978891" lastFinishedPulling="2025-10-05 06:57:42.167627904 +0000 UTC m=+296.050254364" observedRunningTime="2025-10-05 06:57:42.805466925 +0000 UTC m=+296.688093465" watchObservedRunningTime="2025-10-05 06:57:42.808343596 +0000 UTC m=+296.690970056" Oct 05 06:57:43 crc kubenswrapper[4935]: I1005 06:57:43.770941 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbnlj" event={"ID":"d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742","Type":"ContainerStarted","Data":"6cdf0cd2505e0d1125e4aaeeb2d9c7caf81248d7f1b4de6ed85adb9e2a0b9251"} Oct 05 06:57:43 crc kubenswrapper[4935]: I1005 06:57:43.791598 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gbnlj" podStartSLOduration=2.191931012 podStartE2EDuration="4.791576423s" podCreationTimestamp="2025-10-05 06:57:39 +0000 UTC" firstStartedPulling="2025-10-05 06:57:40.735679094 +0000 UTC m=+294.618305554" lastFinishedPulling="2025-10-05 06:57:43.335324515 +0000 UTC m=+297.217950965" observedRunningTime="2025-10-05 06:57:43.791119158 +0000 UTC m=+297.673745618" watchObservedRunningTime="2025-10-05 06:57:43.791576423 +0000 UTC m=+297.674202883" Oct 05 06:57:47 crc kubenswrapper[4935]: I1005 06:57:47.033252 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:47 crc kubenswrapper[4935]: I1005 06:57:47.035165 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:47 crc kubenswrapper[4935]: I1005 06:57:47.090453 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:47 crc kubenswrapper[4935]: I1005 06:57:47.227223 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:47 crc kubenswrapper[4935]: I1005 06:57:47.227488 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:47 crc kubenswrapper[4935]: I1005 06:57:47.279710 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:47 crc kubenswrapper[4935]: I1005 06:57:47.842994 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dqf5x" Oct 05 06:57:47 crc kubenswrapper[4935]: I1005 06:57:47.858754 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cl4vw" Oct 05 06:57:49 crc kubenswrapper[4935]: I1005 06:57:49.428400 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:49 crc kubenswrapper[4935]: I1005 06:57:49.430096 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:49 crc kubenswrapper[4935]: I1005 06:57:49.477028 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:49 crc kubenswrapper[4935]: I1005 06:57:49.648549 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:49 crc kubenswrapper[4935]: I1005 06:57:49.649003 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:49 crc kubenswrapper[4935]: I1005 06:57:49.708974 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:57:49 crc kubenswrapper[4935]: I1005 06:57:49.840741 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gbnlj" Oct 05 06:57:49 crc kubenswrapper[4935]: I1005 06:57:49.841754 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s65n7" Oct 05 06:58:14 crc kubenswrapper[4935]: I1005 06:58:14.289878 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:58:14 crc kubenswrapper[4935]: I1005 06:58:14.290631 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:58:44 crc kubenswrapper[4935]: I1005 06:58:44.289818 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:58:44 crc kubenswrapper[4935]: I1005 06:58:44.290545 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:59:14 crc kubenswrapper[4935]: I1005 06:59:14.289866 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:59:14 crc kubenswrapper[4935]: I1005 06:59:14.290980 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:59:14 crc kubenswrapper[4935]: I1005 06:59:14.291116 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 06:59:14 crc kubenswrapper[4935]: I1005 06:59:14.292148 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f8f2033fd55c6c815714e0c27ced3951dd73a6a07baa92041ec8b26931c0625"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 06:59:14 crc kubenswrapper[4935]: I1005 06:59:14.292213 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://1f8f2033fd55c6c815714e0c27ced3951dd73a6a07baa92041ec8b26931c0625" gracePeriod=600 Oct 05 06:59:15 crc kubenswrapper[4935]: I1005 06:59:15.345804 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="1f8f2033fd55c6c815714e0c27ced3951dd73a6a07baa92041ec8b26931c0625" exitCode=0 Oct 05 06:59:15 crc kubenswrapper[4935]: I1005 06:59:15.345935 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"1f8f2033fd55c6c815714e0c27ced3951dd73a6a07baa92041ec8b26931c0625"} Oct 05 06:59:15 crc kubenswrapper[4935]: I1005 06:59:15.346166 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"875b37b1a2f484d7d3a568e3f0b5abb49b17d16349a6fba870162dc27f30e9b1"} Oct 05 06:59:15 crc kubenswrapper[4935]: I1005 06:59:15.346189 4935 scope.go:117] "RemoveContainer" containerID="7895a31298725bf29dd2256463850004a3eef923fbd96666a597cd291e55d2f8" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.157943 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q"] Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.159997 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.163655 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.164451 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.173391 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q"] Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.284735 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff3c26d6-6c89-4636-a105-33885e9a9de4-config-volume\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.284992 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff3c26d6-6c89-4636-a105-33885e9a9de4-secret-volume\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.285032 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wnt6\" (UniqueName: \"kubernetes.io/projected/ff3c26d6-6c89-4636-a105-33885e9a9de4-kube-api-access-4wnt6\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.385999 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wnt6\" (UniqueName: \"kubernetes.io/projected/ff3c26d6-6c89-4636-a105-33885e9a9de4-kube-api-access-4wnt6\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.386102 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff3c26d6-6c89-4636-a105-33885e9a9de4-config-volume\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.386208 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff3c26d6-6c89-4636-a105-33885e9a9de4-secret-volume\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.387122 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff3c26d6-6c89-4636-a105-33885e9a9de4-config-volume\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.400473 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff3c26d6-6c89-4636-a105-33885e9a9de4-secret-volume\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.408022 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wnt6\" (UniqueName: \"kubernetes.io/projected/ff3c26d6-6c89-4636-a105-33885e9a9de4-kube-api-access-4wnt6\") pod \"collect-profiles-29327460-4km5q\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.489590 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:00 crc kubenswrapper[4935]: I1005 07:00:00.665528 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q"] Oct 05 07:00:01 crc kubenswrapper[4935]: I1005 07:00:01.671831 4935 generic.go:334] "Generic (PLEG): container finished" podID="ff3c26d6-6c89-4636-a105-33885e9a9de4" containerID="8203d88f3096ded755782c14c1ae97226db565f9e28b7cb65b400adb68d28f69" exitCode=0 Oct 05 07:00:01 crc kubenswrapper[4935]: I1005 07:00:01.671911 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" event={"ID":"ff3c26d6-6c89-4636-a105-33885e9a9de4","Type":"ContainerDied","Data":"8203d88f3096ded755782c14c1ae97226db565f9e28b7cb65b400adb68d28f69"} Oct 05 07:00:01 crc kubenswrapper[4935]: I1005 07:00:01.672142 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" event={"ID":"ff3c26d6-6c89-4636-a105-33885e9a9de4","Type":"ContainerStarted","Data":"1b795f1c7ed498a19bd276b7511929990a4ab0be38d0417564c113526c0135b3"} Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.037417 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.224102 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff3c26d6-6c89-4636-a105-33885e9a9de4-secret-volume\") pod \"ff3c26d6-6c89-4636-a105-33885e9a9de4\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.224259 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff3c26d6-6c89-4636-a105-33885e9a9de4-config-volume\") pod \"ff3c26d6-6c89-4636-a105-33885e9a9de4\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.224323 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wnt6\" (UniqueName: \"kubernetes.io/projected/ff3c26d6-6c89-4636-a105-33885e9a9de4-kube-api-access-4wnt6\") pod \"ff3c26d6-6c89-4636-a105-33885e9a9de4\" (UID: \"ff3c26d6-6c89-4636-a105-33885e9a9de4\") " Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.225508 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3c26d6-6c89-4636-a105-33885e9a9de4-config-volume" (OuterVolumeSpecName: "config-volume") pod "ff3c26d6-6c89-4636-a105-33885e9a9de4" (UID: "ff3c26d6-6c89-4636-a105-33885e9a9de4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.232817 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff3c26d6-6c89-4636-a105-33885e9a9de4-kube-api-access-4wnt6" (OuterVolumeSpecName: "kube-api-access-4wnt6") pod "ff3c26d6-6c89-4636-a105-33885e9a9de4" (UID: "ff3c26d6-6c89-4636-a105-33885e9a9de4"). InnerVolumeSpecName "kube-api-access-4wnt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.233002 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff3c26d6-6c89-4636-a105-33885e9a9de4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ff3c26d6-6c89-4636-a105-33885e9a9de4" (UID: "ff3c26d6-6c89-4636-a105-33885e9a9de4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.326332 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff3c26d6-6c89-4636-a105-33885e9a9de4-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.326393 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff3c26d6-6c89-4636-a105-33885e9a9de4-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.326414 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wnt6\" (UniqueName: \"kubernetes.io/projected/ff3c26d6-6c89-4636-a105-33885e9a9de4-kube-api-access-4wnt6\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.688990 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" event={"ID":"ff3c26d6-6c89-4636-a105-33885e9a9de4","Type":"ContainerDied","Data":"1b795f1c7ed498a19bd276b7511929990a4ab0be38d0417564c113526c0135b3"} Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.689063 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q" Oct 05 07:00:03 crc kubenswrapper[4935]: I1005 07:00:03.689064 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b795f1c7ed498a19bd276b7511929990a4ab0be38d0417564c113526c0135b3" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.916810 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8vwqf"] Oct 05 07:00:52 crc kubenswrapper[4935]: E1005 07:00:52.917612 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3c26d6-6c89-4636-a105-33885e9a9de4" containerName="collect-profiles" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.917629 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3c26d6-6c89-4636-a105-33885e9a9de4" containerName="collect-profiles" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.917793 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff3c26d6-6c89-4636-a105-33885e9a9de4" containerName="collect-profiles" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.918310 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.937843 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5dddde27-f96a-42a3-b030-ea56be419b7d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.937921 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-registry-tls\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.937943 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dddde27-f96a-42a3-b030-ea56be419b7d-trusted-ca\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.937979 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5dddde27-f96a-42a3-b030-ea56be419b7d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.937999 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5dddde27-f96a-42a3-b030-ea56be419b7d-registry-certificates\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.938017 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk2fm\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-kube-api-access-xk2fm\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.938042 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.938088 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-bound-sa-token\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.938646 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8vwqf"] Oct 05 07:00:52 crc kubenswrapper[4935]: I1005 07:00:52.975035 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.039366 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-registry-tls\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.039414 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dddde27-f96a-42a3-b030-ea56be419b7d-trusted-ca\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.039445 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5dddde27-f96a-42a3-b030-ea56be419b7d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.039465 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5dddde27-f96a-42a3-b030-ea56be419b7d-registry-certificates\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.039489 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk2fm\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-kube-api-access-xk2fm\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.039521 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-bound-sa-token\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.039552 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5dddde27-f96a-42a3-b030-ea56be419b7d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.040027 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5dddde27-f96a-42a3-b030-ea56be419b7d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.041062 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5dddde27-f96a-42a3-b030-ea56be419b7d-registry-certificates\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.043258 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dddde27-f96a-42a3-b030-ea56be419b7d-trusted-ca\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.045536 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5dddde27-f96a-42a3-b030-ea56be419b7d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.045953 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-registry-tls\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.066419 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-bound-sa-token\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.066482 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk2fm\" (UniqueName: \"kubernetes.io/projected/5dddde27-f96a-42a3-b030-ea56be419b7d-kube-api-access-xk2fm\") pod \"image-registry-66df7c8f76-8vwqf\" (UID: \"5dddde27-f96a-42a3-b030-ea56be419b7d\") " pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.240377 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:53 crc kubenswrapper[4935]: I1005 07:00:53.480954 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8vwqf"] Oct 05 07:00:54 crc kubenswrapper[4935]: I1005 07:00:54.053712 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" event={"ID":"5dddde27-f96a-42a3-b030-ea56be419b7d","Type":"ContainerStarted","Data":"bc57ca785e7b427621104909b1e7022acfa186e62083495419cc40c4a425d780"} Oct 05 07:00:54 crc kubenswrapper[4935]: I1005 07:00:54.054071 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" event={"ID":"5dddde27-f96a-42a3-b030-ea56be419b7d","Type":"ContainerStarted","Data":"82d2095f2322dc1b6154af135aa946ba02743c8a8234d660c9c0fc367c89218c"} Oct 05 07:00:54 crc kubenswrapper[4935]: I1005 07:00:54.054097 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:00:54 crc kubenswrapper[4935]: I1005 07:00:54.078944 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" podStartSLOduration=2.078923212 podStartE2EDuration="2.078923212s" podCreationTimestamp="2025-10-05 07:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:00:54.075266994 +0000 UTC m=+487.957893494" watchObservedRunningTime="2025-10-05 07:00:54.078923212 +0000 UTC m=+487.961549672" Oct 05 07:01:13 crc kubenswrapper[4935]: I1005 07:01:13.248580 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8vwqf" Oct 05 07:01:13 crc kubenswrapper[4935]: I1005 07:01:13.315595 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5skck"] Oct 05 07:01:14 crc kubenswrapper[4935]: I1005 07:01:14.289562 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:01:14 crc kubenswrapper[4935]: I1005 07:01:14.290630 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.365225 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" podUID="a34ebce5-75e3-4033-9fa0-bb17b2151fae" containerName="registry" containerID="cri-o://4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734" gracePeriod=30 Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.748207 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.903316 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a34ebce5-75e3-4033-9fa0-bb17b2151fae-ca-trust-extracted\") pod \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.903477 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.903511 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-certificates\") pod \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.903536 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-bound-sa-token\") pod \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.903568 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gxhl\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-kube-api-access-5gxhl\") pod \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.903597 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-tls\") pod \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.903626 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a34ebce5-75e3-4033-9fa0-bb17b2151fae-installation-pull-secrets\") pod \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.903652 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-trusted-ca\") pod \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\" (UID: \"a34ebce5-75e3-4033-9fa0-bb17b2151fae\") " Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.904669 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a34ebce5-75e3-4033-9fa0-bb17b2151fae" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.905046 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a34ebce5-75e3-4033-9fa0-bb17b2151fae" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.916084 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a34ebce5-75e3-4033-9fa0-bb17b2151fae" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.916483 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-kube-api-access-5gxhl" (OuterVolumeSpecName: "kube-api-access-5gxhl") pod "a34ebce5-75e3-4033-9fa0-bb17b2151fae" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae"). InnerVolumeSpecName "kube-api-access-5gxhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.916567 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a34ebce5-75e3-4033-9fa0-bb17b2151fae-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a34ebce5-75e3-4033-9fa0-bb17b2151fae" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.916644 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a34ebce5-75e3-4033-9fa0-bb17b2151fae" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.931632 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a34ebce5-75e3-4033-9fa0-bb17b2151fae-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a34ebce5-75e3-4033-9fa0-bb17b2151fae" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:38 crc kubenswrapper[4935]: I1005 07:01:38.933574 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a34ebce5-75e3-4033-9fa0-bb17b2151fae" (UID: "a34ebce5-75e3-4033-9fa0-bb17b2151fae"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.004704 4935 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a34ebce5-75e3-4033-9fa0-bb17b2151fae-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.004753 4935 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.004776 4935 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.004791 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gxhl\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-kube-api-access-5gxhl\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.004809 4935 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a34ebce5-75e3-4033-9fa0-bb17b2151fae-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.004824 4935 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a34ebce5-75e3-4033-9fa0-bb17b2151fae-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.004838 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a34ebce5-75e3-4033-9fa0-bb17b2151fae-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.344390 4935 generic.go:334] "Generic (PLEG): container finished" podID="a34ebce5-75e3-4033-9fa0-bb17b2151fae" containerID="4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734" exitCode=0 Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.344469 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.344481 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" event={"ID":"a34ebce5-75e3-4033-9fa0-bb17b2151fae","Type":"ContainerDied","Data":"4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734"} Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.344565 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5skck" event={"ID":"a34ebce5-75e3-4033-9fa0-bb17b2151fae","Type":"ContainerDied","Data":"ad3e5031b0fd2956c9a8fc76d178d71234eeffb01931b759e0f25fc2d137799d"} Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.344615 4935 scope.go:117] "RemoveContainer" containerID="4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.370319 4935 scope.go:117] "RemoveContainer" containerID="4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734" Oct 05 07:01:39 crc kubenswrapper[4935]: E1005 07:01:39.372098 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734\": container with ID starting with 4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734 not found: ID does not exist" containerID="4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.372151 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734"} err="failed to get container status \"4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734\": rpc error: code = NotFound desc = could not find container \"4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734\": container with ID starting with 4be07d778cad51b1513951c5a29e9f539359701625847964c1ca38375b34e734 not found: ID does not exist" Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.385298 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5skck"] Oct 05 07:01:39 crc kubenswrapper[4935]: I1005 07:01:39.388553 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5skck"] Oct 05 07:01:40 crc kubenswrapper[4935]: I1005 07:01:40.783344 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a34ebce5-75e3-4033-9fa0-bb17b2151fae" path="/var/lib/kubelet/pods/a34ebce5-75e3-4033-9fa0-bb17b2151fae/volumes" Oct 05 07:01:44 crc kubenswrapper[4935]: I1005 07:01:44.289487 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:01:44 crc kubenswrapper[4935]: I1005 07:01:44.289945 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:02:14 crc kubenswrapper[4935]: I1005 07:02:14.290118 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:02:14 crc kubenswrapper[4935]: I1005 07:02:14.291300 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:02:14 crc kubenswrapper[4935]: I1005 07:02:14.291392 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:02:14 crc kubenswrapper[4935]: I1005 07:02:14.292212 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"875b37b1a2f484d7d3a568e3f0b5abb49b17d16349a6fba870162dc27f30e9b1"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:02:14 crc kubenswrapper[4935]: I1005 07:02:14.292270 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://875b37b1a2f484d7d3a568e3f0b5abb49b17d16349a6fba870162dc27f30e9b1" gracePeriod=600 Oct 05 07:02:14 crc kubenswrapper[4935]: I1005 07:02:14.567665 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="875b37b1a2f484d7d3a568e3f0b5abb49b17d16349a6fba870162dc27f30e9b1" exitCode=0 Oct 05 07:02:14 crc kubenswrapper[4935]: I1005 07:02:14.567762 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"875b37b1a2f484d7d3a568e3f0b5abb49b17d16349a6fba870162dc27f30e9b1"} Oct 05 07:02:14 crc kubenswrapper[4935]: I1005 07:02:14.568087 4935 scope.go:117] "RemoveContainer" containerID="1f8f2033fd55c6c815714e0c27ced3951dd73a6a07baa92041ec8b26931c0625" Oct 05 07:02:15 crc kubenswrapper[4935]: I1005 07:02:15.573868 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"16ac36f74f68eb8c2c9e34c4f313b94ebe39f5ac0b6b53af3d6a75adcbd15fbc"} Oct 05 07:04:14 crc kubenswrapper[4935]: I1005 07:04:14.289376 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:04:14 crc kubenswrapper[4935]: I1005 07:04:14.289944 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:04:44 crc kubenswrapper[4935]: I1005 07:04:44.289751 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:04:44 crc kubenswrapper[4935]: I1005 07:04:44.290494 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.289295 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.290137 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.290212 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.291219 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"16ac36f74f68eb8c2c9e34c4f313b94ebe39f5ac0b6b53af3d6a75adcbd15fbc"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.291309 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://16ac36f74f68eb8c2c9e34c4f313b94ebe39f5ac0b6b53af3d6a75adcbd15fbc" gracePeriod=600 Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.671928 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="16ac36f74f68eb8c2c9e34c4f313b94ebe39f5ac0b6b53af3d6a75adcbd15fbc" exitCode=0 Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.672015 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"16ac36f74f68eb8c2c9e34c4f313b94ebe39f5ac0b6b53af3d6a75adcbd15fbc"} Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.672284 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"52b267e7f2c1f92355886ddbf4cef30a218042be6e418507f266c72f41da3c70"} Oct 05 07:05:14 crc kubenswrapper[4935]: I1005 07:05:14.672306 4935 scope.go:117] "RemoveContainer" containerID="875b37b1a2f484d7d3a568e3f0b5abb49b17d16349a6fba870162dc27f30e9b1" Oct 05 07:05:17 crc kubenswrapper[4935]: I1005 07:05:17.971449 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nh99w"] Oct 05 07:05:17 crc kubenswrapper[4935]: I1005 07:05:17.972607 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" podUID="4fb78a87-5385-4785-b4b4-4e077d5000ed" containerName="controller-manager" containerID="cri-o://2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890" gracePeriod=30 Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.066392 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7"] Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.066643 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" podUID="d90f77af-0ce4-4b2f-87d2-049e45327404" containerName="route-controller-manager" containerID="cri-o://068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82" gracePeriod=30 Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.329987 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.410840 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.442367 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-config\") pod \"4fb78a87-5385-4785-b4b4-4e077d5000ed\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.442442 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fb78a87-5385-4785-b4b4-4e077d5000ed-serving-cert\") pod \"4fb78a87-5385-4785-b4b4-4e077d5000ed\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.442464 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-client-ca\") pod \"4fb78a87-5385-4785-b4b4-4e077d5000ed\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.442503 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-proxy-ca-bundles\") pod \"4fb78a87-5385-4785-b4b4-4e077d5000ed\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.442627 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bgr8\" (UniqueName: \"kubernetes.io/projected/4fb78a87-5385-4785-b4b4-4e077d5000ed-kube-api-access-9bgr8\") pod \"4fb78a87-5385-4785-b4b4-4e077d5000ed\" (UID: \"4fb78a87-5385-4785-b4b4-4e077d5000ed\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.443278 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-client-ca" (OuterVolumeSpecName: "client-ca") pod "4fb78a87-5385-4785-b4b4-4e077d5000ed" (UID: "4fb78a87-5385-4785-b4b4-4e077d5000ed"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.443287 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4fb78a87-5385-4785-b4b4-4e077d5000ed" (UID: "4fb78a87-5385-4785-b4b4-4e077d5000ed"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.443505 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-config" (OuterVolumeSpecName: "config") pod "4fb78a87-5385-4785-b4b4-4e077d5000ed" (UID: "4fb78a87-5385-4785-b4b4-4e077d5000ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.448962 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fb78a87-5385-4785-b4b4-4e077d5000ed-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4fb78a87-5385-4785-b4b4-4e077d5000ed" (UID: "4fb78a87-5385-4785-b4b4-4e077d5000ed"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.449877 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fb78a87-5385-4785-b4b4-4e077d5000ed-kube-api-access-9bgr8" (OuterVolumeSpecName: "kube-api-access-9bgr8") pod "4fb78a87-5385-4785-b4b4-4e077d5000ed" (UID: "4fb78a87-5385-4785-b4b4-4e077d5000ed"). InnerVolumeSpecName "kube-api-access-9bgr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.543581 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-client-ca\") pod \"d90f77af-0ce4-4b2f-87d2-049e45327404\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.543711 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rr7r\" (UniqueName: \"kubernetes.io/projected/d90f77af-0ce4-4b2f-87d2-049e45327404-kube-api-access-8rr7r\") pod \"d90f77af-0ce4-4b2f-87d2-049e45327404\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.543752 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-config\") pod \"d90f77af-0ce4-4b2f-87d2-049e45327404\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.543796 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d90f77af-0ce4-4b2f-87d2-049e45327404-serving-cert\") pod \"d90f77af-0ce4-4b2f-87d2-049e45327404\" (UID: \"d90f77af-0ce4-4b2f-87d2-049e45327404\") " Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.543974 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.543989 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fb78a87-5385-4785-b4b4-4e077d5000ed-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.543998 4935 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.544008 4935 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4fb78a87-5385-4785-b4b4-4e077d5000ed-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.544019 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bgr8\" (UniqueName: \"kubernetes.io/projected/4fb78a87-5385-4785-b4b4-4e077d5000ed-kube-api-access-9bgr8\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.544243 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-client-ca" (OuterVolumeSpecName: "client-ca") pod "d90f77af-0ce4-4b2f-87d2-049e45327404" (UID: "d90f77af-0ce4-4b2f-87d2-049e45327404"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.545236 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-config" (OuterVolumeSpecName: "config") pod "d90f77af-0ce4-4b2f-87d2-049e45327404" (UID: "d90f77af-0ce4-4b2f-87d2-049e45327404"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.549338 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d90f77af-0ce4-4b2f-87d2-049e45327404-kube-api-access-8rr7r" (OuterVolumeSpecName: "kube-api-access-8rr7r") pod "d90f77af-0ce4-4b2f-87d2-049e45327404" (UID: "d90f77af-0ce4-4b2f-87d2-049e45327404"). InnerVolumeSpecName "kube-api-access-8rr7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.549663 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d90f77af-0ce4-4b2f-87d2-049e45327404-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d90f77af-0ce4-4b2f-87d2-049e45327404" (UID: "d90f77af-0ce4-4b2f-87d2-049e45327404"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.645189 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d90f77af-0ce4-4b2f-87d2-049e45327404-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.645223 4935 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.645233 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rr7r\" (UniqueName: \"kubernetes.io/projected/d90f77af-0ce4-4b2f-87d2-049e45327404-kube-api-access-8rr7r\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.645242 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d90f77af-0ce4-4b2f-87d2-049e45327404-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.700204 4935 generic.go:334] "Generic (PLEG): container finished" podID="4fb78a87-5385-4785-b4b4-4e077d5000ed" containerID="2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890" exitCode=0 Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.700279 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" event={"ID":"4fb78a87-5385-4785-b4b4-4e077d5000ed","Type":"ContainerDied","Data":"2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890"} Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.700306 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" event={"ID":"4fb78a87-5385-4785-b4b4-4e077d5000ed","Type":"ContainerDied","Data":"d3de02116d55c7f1b4eda70beff2e4443e2e247134aa171eb917401adbf0390e"} Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.700322 4935 scope.go:117] "RemoveContainer" containerID="2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.700756 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nh99w" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.705345 4935 generic.go:334] "Generic (PLEG): container finished" podID="d90f77af-0ce4-4b2f-87d2-049e45327404" containerID="068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82" exitCode=0 Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.705381 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" event={"ID":"d90f77af-0ce4-4b2f-87d2-049e45327404","Type":"ContainerDied","Data":"068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82"} Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.705412 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" event={"ID":"d90f77af-0ce4-4b2f-87d2-049e45327404","Type":"ContainerDied","Data":"69c3c2cc36de60ce4ea8294fa1bf47b0397e81a0d64a1cd7cb3529d39d08279b"} Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.705472 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.720320 4935 scope.go:117] "RemoveContainer" containerID="2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890" Oct 05 07:05:18 crc kubenswrapper[4935]: E1005 07:05:18.720771 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890\": container with ID starting with 2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890 not found: ID does not exist" containerID="2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.720824 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890"} err="failed to get container status \"2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890\": rpc error: code = NotFound desc = could not find container \"2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890\": container with ID starting with 2a365920e00110b27dcd1d3c976f735e0d5cbc44e1ca29bda949d099f873e890 not found: ID does not exist" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.720856 4935 scope.go:117] "RemoveContainer" containerID="068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.739510 4935 scope.go:117] "RemoveContainer" containerID="068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82" Oct 05 07:05:18 crc kubenswrapper[4935]: E1005 07:05:18.740251 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82\": container with ID starting with 068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82 not found: ID does not exist" containerID="068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.740566 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82"} err="failed to get container status \"068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82\": rpc error: code = NotFound desc = could not find container \"068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82\": container with ID starting with 068e297353831dee7ab21e2386fd03bb3abf8ff56185e506cbdd1e03bcf3de82 not found: ID does not exist" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.757675 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7"] Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.765946 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwzq7"] Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.773349 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nh99w"] Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.788426 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d90f77af-0ce4-4b2f-87d2-049e45327404" path="/var/lib/kubelet/pods/d90f77af-0ce4-4b2f-87d2-049e45327404/volumes" Oct 05 07:05:18 crc kubenswrapper[4935]: I1005 07:05:18.789272 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nh99w"] Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.739366 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl"] Oct 05 07:05:19 crc kubenswrapper[4935]: E1005 07:05:19.740037 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34ebce5-75e3-4033-9fa0-bb17b2151fae" containerName="registry" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.740052 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34ebce5-75e3-4033-9fa0-bb17b2151fae" containerName="registry" Oct 05 07:05:19 crc kubenswrapper[4935]: E1005 07:05:19.740067 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d90f77af-0ce4-4b2f-87d2-049e45327404" containerName="route-controller-manager" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.740075 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d90f77af-0ce4-4b2f-87d2-049e45327404" containerName="route-controller-manager" Oct 05 07:05:19 crc kubenswrapper[4935]: E1005 07:05:19.740096 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fb78a87-5385-4785-b4b4-4e077d5000ed" containerName="controller-manager" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.740105 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fb78a87-5385-4785-b4b4-4e077d5000ed" containerName="controller-manager" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.740217 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a34ebce5-75e3-4033-9fa0-bb17b2151fae" containerName="registry" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.740228 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d90f77af-0ce4-4b2f-87d2-049e45327404" containerName="route-controller-manager" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.740242 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fb78a87-5385-4785-b4b4-4e077d5000ed" containerName="controller-manager" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.740613 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.744336 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.744446 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.744518 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.744820 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.745348 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.745344 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl"] Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.746075 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.749552 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.750391 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.751289 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.751333 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.751340 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.751794 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.751806 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.755775 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl"] Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.765057 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.769786 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl"] Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870106 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-serving-cert\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870184 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-client-ca\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870273 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-client-ca\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870301 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7822b3a4-9caa-4dde-94cc-523a5f213d97-serving-cert\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870322 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-proxy-ca-bundles\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870353 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slhxt\" (UniqueName: \"kubernetes.io/projected/7822b3a4-9caa-4dde-94cc-523a5f213d97-kube-api-access-slhxt\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870378 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-config\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870398 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-config\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.870480 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lgdc\" (UniqueName: \"kubernetes.io/projected/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-kube-api-access-2lgdc\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.951139 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl"] Oct 05 07:05:19 crc kubenswrapper[4935]: E1005 07:05:19.951632 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-slhxt serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" podUID="7822b3a4-9caa-4dde-94cc-523a5f213d97" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.971737 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lgdc\" (UniqueName: \"kubernetes.io/projected/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-kube-api-access-2lgdc\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.971824 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-serving-cert\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.971854 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-client-ca\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.971873 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-client-ca\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.971912 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7822b3a4-9caa-4dde-94cc-523a5f213d97-serving-cert\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.971937 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-proxy-ca-bundles\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.971964 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slhxt\" (UniqueName: \"kubernetes.io/projected/7822b3a4-9caa-4dde-94cc-523a5f213d97-kube-api-access-slhxt\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.971988 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-config\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.972015 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-config\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.973409 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-client-ca\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.973413 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-client-ca\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.973598 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-config\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.974023 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-proxy-ca-bundles\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.974717 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-config\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.978298 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7822b3a4-9caa-4dde-94cc-523a5f213d97-serving-cert\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.989339 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-serving-cert\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.993618 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slhxt\" (UniqueName: \"kubernetes.io/projected/7822b3a4-9caa-4dde-94cc-523a5f213d97-kube-api-access-slhxt\") pod \"route-controller-manager-598d844d48-cg9bl\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:19 crc kubenswrapper[4935]: I1005 07:05:19.993659 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lgdc\" (UniqueName: \"kubernetes.io/projected/4bcf8f95-d8a9-425c-9f40-02ede72c4dc5-kube-api-access-2lgdc\") pod \"controller-manager-794b7ff6cb-cjwdl\" (UID: \"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5\") " pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.065700 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.337901 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl"] Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.721356 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.721379 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" event={"ID":"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5","Type":"ContainerStarted","Data":"4f6e52767e169dc24a57fab9699d4394b006d286cc448423339eba24377285ff"} Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.722037 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" event={"ID":"4bcf8f95-d8a9-425c-9f40-02ede72c4dc5","Type":"ContainerStarted","Data":"cc4ac672365a52ab9b8fe29204739ffc057a9a4128299332109a25fed397f12f"} Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.733287 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.787658 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fb78a87-5385-4785-b4b4-4e077d5000ed" path="/var/lib/kubelet/pods/4fb78a87-5385-4785-b4b4-4e077d5000ed/volumes" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.884780 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7822b3a4-9caa-4dde-94cc-523a5f213d97-serving-cert\") pod \"7822b3a4-9caa-4dde-94cc-523a5f213d97\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.884864 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-config\") pod \"7822b3a4-9caa-4dde-94cc-523a5f213d97\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.884911 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slhxt\" (UniqueName: \"kubernetes.io/projected/7822b3a4-9caa-4dde-94cc-523a5f213d97-kube-api-access-slhxt\") pod \"7822b3a4-9caa-4dde-94cc-523a5f213d97\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.884955 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-client-ca\") pod \"7822b3a4-9caa-4dde-94cc-523a5f213d97\" (UID: \"7822b3a4-9caa-4dde-94cc-523a5f213d97\") " Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.885588 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-client-ca" (OuterVolumeSpecName: "client-ca") pod "7822b3a4-9caa-4dde-94cc-523a5f213d97" (UID: "7822b3a4-9caa-4dde-94cc-523a5f213d97"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.885648 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-config" (OuterVolumeSpecName: "config") pod "7822b3a4-9caa-4dde-94cc-523a5f213d97" (UID: "7822b3a4-9caa-4dde-94cc-523a5f213d97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.894273 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7822b3a4-9caa-4dde-94cc-523a5f213d97-kube-api-access-slhxt" (OuterVolumeSpecName: "kube-api-access-slhxt") pod "7822b3a4-9caa-4dde-94cc-523a5f213d97" (UID: "7822b3a4-9caa-4dde-94cc-523a5f213d97"). InnerVolumeSpecName "kube-api-access-slhxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.894268 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7822b3a4-9caa-4dde-94cc-523a5f213d97-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7822b3a4-9caa-4dde-94cc-523a5f213d97" (UID: "7822b3a4-9caa-4dde-94cc-523a5f213d97"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.986681 4935 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7822b3a4-9caa-4dde-94cc-523a5f213d97-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.986714 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.986724 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slhxt\" (UniqueName: \"kubernetes.io/projected/7822b3a4-9caa-4dde-94cc-523a5f213d97-kube-api-access-slhxt\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:20 crc kubenswrapper[4935]: I1005 07:05:20.986748 4935 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7822b3a4-9caa-4dde-94cc-523a5f213d97-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.727085 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.756426 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" podStartSLOduration=3.756406934 podStartE2EDuration="3.756406934s" podCreationTimestamp="2025-10-05 07:05:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:05:21.754151254 +0000 UTC m=+755.636777724" watchObservedRunningTime="2025-10-05 07:05:21.756406934 +0000 UTC m=+755.639033394" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.800113 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq"] Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.801121 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.809403 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl"] Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.810816 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.811492 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.811537 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.811699 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.812201 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.812570 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598d844d48-cg9bl"] Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.814963 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq"] Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.815161 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.899774 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e9eab65-fde1-4f1c-aba6-3451ea99384e-client-ca\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.899851 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9eab65-fde1-4f1c-aba6-3451ea99384e-config\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.899870 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wqfp\" (UniqueName: \"kubernetes.io/projected/7e9eab65-fde1-4f1c-aba6-3451ea99384e-kube-api-access-8wqfp\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:21 crc kubenswrapper[4935]: I1005 07:05:21.899919 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e9eab65-fde1-4f1c-aba6-3451ea99384e-serving-cert\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.001313 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9eab65-fde1-4f1c-aba6-3451ea99384e-config\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.001360 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wqfp\" (UniqueName: \"kubernetes.io/projected/7e9eab65-fde1-4f1c-aba6-3451ea99384e-kube-api-access-8wqfp\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.001381 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e9eab65-fde1-4f1c-aba6-3451ea99384e-serving-cert\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.001442 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e9eab65-fde1-4f1c-aba6-3451ea99384e-client-ca\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.002272 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e9eab65-fde1-4f1c-aba6-3451ea99384e-client-ca\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.002563 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9eab65-fde1-4f1c-aba6-3451ea99384e-config\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.005837 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e9eab65-fde1-4f1c-aba6-3451ea99384e-serving-cert\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.019312 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wqfp\" (UniqueName: \"kubernetes.io/projected/7e9eab65-fde1-4f1c-aba6-3451ea99384e-kube-api-access-8wqfp\") pod \"route-controller-manager-7b46457668-k8czq\" (UID: \"7e9eab65-fde1-4f1c-aba6-3451ea99384e\") " pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.131074 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.306745 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq"] Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.734684 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" event={"ID":"7e9eab65-fde1-4f1c-aba6-3451ea99384e","Type":"ContainerStarted","Data":"dedd551af1b82d09dd8def379bd65d30017fc10733e4ee2ddd3677dfc1c01726"} Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.735181 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.735200 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" event={"ID":"7e9eab65-fde1-4f1c-aba6-3451ea99384e","Type":"ContainerStarted","Data":"dc73ece1dec7eddaae89527fcd19f16327f1eb496598d562fe70ea887ea9e18d"} Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.756166 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" podStartSLOduration=3.756150649 podStartE2EDuration="3.756150649s" podCreationTimestamp="2025-10-05 07:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:05:22.754936807 +0000 UTC m=+756.637563297" watchObservedRunningTime="2025-10-05 07:05:22.756150649 +0000 UTC m=+756.638777109" Oct 05 07:05:22 crc kubenswrapper[4935]: I1005 07:05:22.783302 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7822b3a4-9caa-4dde-94cc-523a5f213d97" path="/var/lib/kubelet/pods/7822b3a4-9caa-4dde-94cc-523a5f213d97/volumes" Oct 05 07:05:23 crc kubenswrapper[4935]: I1005 07:05:23.113355 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7b46457668-k8czq" Oct 05 07:05:24 crc kubenswrapper[4935]: I1005 07:05:24.387384 4935 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 07:05:30 crc kubenswrapper[4935]: I1005 07:05:30.065976 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:30 crc kubenswrapper[4935]: I1005 07:05:30.071586 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-794b7ff6cb-cjwdl" Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.668386 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ns4c5"] Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.669264 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovn-controller" containerID="cri-o://f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c" gracePeriod=30 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.669577 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="sbdb" containerID="cri-o://3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5" gracePeriod=30 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.669621 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="nbdb" containerID="cri-o://8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b" gracePeriod=30 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.669655 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="northd" containerID="cri-o://b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b" gracePeriod=30 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.669683 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0" gracePeriod=30 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.669708 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kube-rbac-proxy-node" containerID="cri-o://9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e" gracePeriod=30 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.669733 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovn-acl-logging" containerID="cri-o://f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50" gracePeriod=30 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.795737 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" containerID="cri-o://3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc" gracePeriod=30 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.834907 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/2.log" Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.835227 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/1.log" Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.835261 4935 generic.go:334] "Generic (PLEG): container finished" podID="2b9ff491-4f10-4f1b-822b-a8c46871ce60" containerID="d22b4290c09db56259541f5d949e9b59e16b541d40ce02e695c7ba63bcb3961f" exitCode=2 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.835316 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-687xm" event={"ID":"2b9ff491-4f10-4f1b-822b-a8c46871ce60","Type":"ContainerDied","Data":"d22b4290c09db56259541f5d949e9b59e16b541d40ce02e695c7ba63bcb3961f"} Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.835366 4935 scope.go:117] "RemoveContainer" containerID="7bb8642f2b218d735fad4fa41abcc4fd0f73edc2b6f14588afcb3597391ffd79" Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.835836 4935 scope.go:117] "RemoveContainer" containerID="d22b4290c09db56259541f5d949e9b59e16b541d40ce02e695c7ba63bcb3961f" Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.837693 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/3.log" Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.839818 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovn-acl-logging/0.log" Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.840883 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovn-controller/0.log" Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.841306 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0" exitCode=0 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.841323 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e" exitCode=0 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.841333 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50" exitCode=143 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.841339 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c" exitCode=143 Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.841357 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0"} Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.841377 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e"} Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.841387 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50"} Oct 05 07:05:37 crc kubenswrapper[4935]: I1005 07:05:37.841396 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.028721 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/3.log" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.031173 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovn-acl-logging/0.log" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.031653 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovn-controller/0.log" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.032046 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.082195 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pbfk4"] Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.082568 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.082647 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.082709 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.082765 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.082822 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="northd" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.082877 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="northd" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.082960 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kubecfg-setup" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083011 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kubecfg-setup" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.083059 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="nbdb" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083109 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="nbdb" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.083157 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083207 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.083260 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="sbdb" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083310 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="sbdb" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.083363 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovn-acl-logging" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083436 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovn-acl-logging" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.083494 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovn-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083553 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovn-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.083611 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kube-rbac-proxy-node" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083665 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kube-rbac-proxy-node" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.083718 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083770 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.083957 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="nbdb" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084025 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084074 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kube-rbac-proxy-node" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084125 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="sbdb" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084177 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084223 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="northd" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084279 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovn-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084331 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084384 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovn-acl-logging" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084435 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084489 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.084622 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084683 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: E1005 07:05:38.084736 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084784 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.084939 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerName="ovnkube-controller" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.086491 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116765 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-env-overrides\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116806 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-log-socket\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116828 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116844 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-ovn\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116861 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-etc-openvswitch\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116878 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-netns\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116942 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-script-lib\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116982 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-systemd-units\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116976 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.116997 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-var-lib-openvswitch\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117024 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117010 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117069 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117069 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovn-node-metrics-cert\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117055 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117102 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzplz\" (UniqueName: \"kubernetes.io/projected/019e7dde-1935-4b4c-8312-99f52b6d22fc-kube-api-access-nzplz\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117129 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-slash\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117144 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-config\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117171 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-kubelet\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117191 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-bin\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117210 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-systemd\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117225 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-openvswitch\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117257 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-node-log\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117274 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-netd\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117291 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-ovn-kubernetes\") pod \"019e7dde-1935-4b4c-8312-99f52b6d22fc\" (UID: \"019e7dde-1935-4b4c-8312-99f52b6d22fc\") " Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117374 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117408 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117424 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117443 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117468 4935 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117485 4935 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117498 4935 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117508 4935 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117521 4935 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117531 4935 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117542 4935 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117554 4935 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117464 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117493 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-node-log" (OuterVolumeSpecName: "node-log") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117513 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117536 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-slash" (OuterVolumeSpecName: "host-slash") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117878 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117923 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.117944 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.118068 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-log-socket" (OuterVolumeSpecName: "log-socket") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.123116 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/019e7dde-1935-4b4c-8312-99f52b6d22fc-kube-api-access-nzplz" (OuterVolumeSpecName: "kube-api-access-nzplz") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "kube-api-access-nzplz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.123698 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.130423 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "019e7dde-1935-4b4c-8312-99f52b6d22fc" (UID: "019e7dde-1935-4b4c-8312-99f52b6d22fc"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218685 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-env-overrides\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218750 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovnkube-config\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218771 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-cni-netd\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218792 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-log-socket\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218812 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-systemd-units\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218826 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-node-log\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218847 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-etc-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218862 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-cni-bin\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218879 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovnkube-script-lib\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218924 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218944 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-slash\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.218988 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh4nh\" (UniqueName: \"kubernetes.io/projected/52cb3655-07ef-4ad8-8828-53ff2491c68c-kube-api-access-sh4nh\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219009 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219033 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-systemd\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219056 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-ovn\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219073 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovn-node-metrics-cert\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219089 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-kubelet\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219112 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-run-netns\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219136 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-var-lib-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219159 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219204 4935 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219216 4935 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-log-socket\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219227 4935 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219238 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219249 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzplz\" (UniqueName: \"kubernetes.io/projected/019e7dde-1935-4b4c-8312-99f52b6d22fc-kube-api-access-nzplz\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219260 4935 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-slash\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219271 4935 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/019e7dde-1935-4b4c-8312-99f52b6d22fc-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219280 4935 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219289 4935 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219297 4935 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219305 4935 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-node-log\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.219314 4935 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/019e7dde-1935-4b4c-8312-99f52b6d22fc-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320090 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320245 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320343 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-slash\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320385 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-slash\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320423 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh4nh\" (UniqueName: \"kubernetes.io/projected/52cb3655-07ef-4ad8-8828-53ff2491c68c-kube-api-access-sh4nh\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320544 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320582 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-ovn\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320604 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-systemd\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320622 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovn-node-metrics-cert\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320640 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-kubelet\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320676 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-run-netns\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320705 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-var-lib-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320743 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320759 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320804 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-env-overrides\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.320971 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovnkube-config\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321038 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-log-socket\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321091 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-cni-netd\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321172 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-systemd-units\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321219 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-node-log\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321309 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-etc-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321329 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-env-overrides\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321365 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-cni-bin\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321381 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-cni-netd\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321402 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovnkube-script-lib\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321418 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321387 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-run-netns\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321369 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-kubelet\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321445 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-systemd\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321404 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-var-lib-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321534 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-node-log\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321568 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-etc-openvswitch\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321567 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-host-cni-bin\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321591 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-systemd-units\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321601 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-log-socket\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.321629 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/52cb3655-07ef-4ad8-8828-53ff2491c68c-run-ovn\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.322015 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovnkube-config\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.322828 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovnkube-script-lib\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.323708 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52cb3655-07ef-4ad8-8828-53ff2491c68c-ovn-node-metrics-cert\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.343019 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh4nh\" (UniqueName: \"kubernetes.io/projected/52cb3655-07ef-4ad8-8828-53ff2491c68c-kube-api-access-sh4nh\") pod \"ovnkube-node-pbfk4\" (UID: \"52cb3655-07ef-4ad8-8828-53ff2491c68c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.400595 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:38 crc kubenswrapper[4935]: W1005 07:05:38.420362 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52cb3655_07ef_4ad8_8828_53ff2491c68c.slice/crio-a3b01b81c2662df2455ad0daf7a9d8e917db9c97a566ade35dae21ae1fd422b2 WatchSource:0}: Error finding container a3b01b81c2662df2455ad0daf7a9d8e917db9c97a566ade35dae21ae1fd422b2: Status 404 returned error can't find the container with id a3b01b81c2662df2455ad0daf7a9d8e917db9c97a566ade35dae21ae1fd422b2 Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.847676 4935 generic.go:334] "Generic (PLEG): container finished" podID="52cb3655-07ef-4ad8-8828-53ff2491c68c" containerID="e1790606eaee2b711a65949784cecf7873923ec4e3f8bedeaf1fbe0bb198a3dd" exitCode=0 Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.847762 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerDied","Data":"e1790606eaee2b711a65949784cecf7873923ec4e3f8bedeaf1fbe0bb198a3dd"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.848220 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"a3b01b81c2662df2455ad0daf7a9d8e917db9c97a566ade35dae21ae1fd422b2"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.850883 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-687xm_2b9ff491-4f10-4f1b-822b-a8c46871ce60/kube-multus/2.log" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.851062 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-687xm" event={"ID":"2b9ff491-4f10-4f1b-822b-a8c46871ce60","Type":"ContainerStarted","Data":"c14dbb8707e807be06b9675dba05f917cbb3bee22d4b861dc1be9984a84125e1"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.859037 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovnkube-controller/3.log" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.863242 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovn-acl-logging/0.log" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.865134 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ns4c5_019e7dde-1935-4b4c-8312-99f52b6d22fc/ovn-controller/0.log" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866035 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc" exitCode=0 Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866090 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5" exitCode=0 Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866103 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b" exitCode=0 Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866110 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866136 4935 generic.go:334] "Generic (PLEG): container finished" podID="019e7dde-1935-4b4c-8312-99f52b6d22fc" containerID="b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b" exitCode=0 Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866178 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866203 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866221 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866240 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" event={"ID":"019e7dde-1935-4b4c-8312-99f52b6d22fc","Type":"ContainerDied","Data":"02438eba25ad62d4e9a84e0e6b032ee5db5a92a64239d6a9e862635d9cbb9442"} Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866248 4935 scope.go:117] "RemoveContainer" containerID="3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.866604 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ns4c5" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.886060 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.920990 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ns4c5"] Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.921424 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ns4c5"] Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.929036 4935 scope.go:117] "RemoveContainer" containerID="3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.949320 4935 scope.go:117] "RemoveContainer" containerID="8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.973383 4935 scope.go:117] "RemoveContainer" containerID="b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b" Oct 05 07:05:38 crc kubenswrapper[4935]: I1005 07:05:38.992878 4935 scope.go:117] "RemoveContainer" containerID="aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.006709 4935 scope.go:117] "RemoveContainer" containerID="9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.022770 4935 scope.go:117] "RemoveContainer" containerID="f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.039970 4935 scope.go:117] "RemoveContainer" containerID="f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.076717 4935 scope.go:117] "RemoveContainer" containerID="80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.114328 4935 scope.go:117] "RemoveContainer" containerID="3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.114747 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": container with ID starting with 3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc not found: ID does not exist" containerID="3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.114786 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc"} err="failed to get container status \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": rpc error: code = NotFound desc = could not find container \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": container with ID starting with 3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.114810 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.115191 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": container with ID starting with 1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b not found: ID does not exist" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.115254 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b"} err="failed to get container status \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": rpc error: code = NotFound desc = could not find container \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": container with ID starting with 1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.115296 4935 scope.go:117] "RemoveContainer" containerID="3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.115675 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": container with ID starting with 3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5 not found: ID does not exist" containerID="3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.115706 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5"} err="failed to get container status \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": rpc error: code = NotFound desc = could not find container \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": container with ID starting with 3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.115727 4935 scope.go:117] "RemoveContainer" containerID="8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.115974 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": container with ID starting with 8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b not found: ID does not exist" containerID="8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.115996 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b"} err="failed to get container status \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": rpc error: code = NotFound desc = could not find container \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": container with ID starting with 8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.116011 4935 scope.go:117] "RemoveContainer" containerID="b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.116593 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": container with ID starting with b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b not found: ID does not exist" containerID="b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.116657 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b"} err="failed to get container status \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": rpc error: code = NotFound desc = could not find container \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": container with ID starting with b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.116679 4935 scope.go:117] "RemoveContainer" containerID="aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.117011 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": container with ID starting with aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0 not found: ID does not exist" containerID="aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.117064 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0"} err="failed to get container status \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": rpc error: code = NotFound desc = could not find container \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": container with ID starting with aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.117083 4935 scope.go:117] "RemoveContainer" containerID="9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.117292 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": container with ID starting with 9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e not found: ID does not exist" containerID="9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.117313 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e"} err="failed to get container status \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": rpc error: code = NotFound desc = could not find container \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": container with ID starting with 9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.117330 4935 scope.go:117] "RemoveContainer" containerID="f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.117553 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": container with ID starting with f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50 not found: ID does not exist" containerID="f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.117575 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50"} err="failed to get container status \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": rpc error: code = NotFound desc = could not find container \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": container with ID starting with f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.117592 4935 scope.go:117] "RemoveContainer" containerID="f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.117934 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": container with ID starting with f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c not found: ID does not exist" containerID="f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.117958 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c"} err="failed to get container status \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": rpc error: code = NotFound desc = could not find container \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": container with ID starting with f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.117975 4935 scope.go:117] "RemoveContainer" containerID="80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa" Oct 05 07:05:39 crc kubenswrapper[4935]: E1005 07:05:39.118204 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": container with ID starting with 80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa not found: ID does not exist" containerID="80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.118226 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa"} err="failed to get container status \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": rpc error: code = NotFound desc = could not find container \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": container with ID starting with 80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.118244 4935 scope.go:117] "RemoveContainer" containerID="3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.118532 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc"} err="failed to get container status \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": rpc error: code = NotFound desc = could not find container \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": container with ID starting with 3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.118550 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.118838 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b"} err="failed to get container status \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": rpc error: code = NotFound desc = could not find container \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": container with ID starting with 1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.118857 4935 scope.go:117] "RemoveContainer" containerID="3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.119182 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5"} err="failed to get container status \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": rpc error: code = NotFound desc = could not find container \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": container with ID starting with 3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.119204 4935 scope.go:117] "RemoveContainer" containerID="8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.119471 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b"} err="failed to get container status \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": rpc error: code = NotFound desc = could not find container \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": container with ID starting with 8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.119490 4935 scope.go:117] "RemoveContainer" containerID="b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.119678 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b"} err="failed to get container status \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": rpc error: code = NotFound desc = could not find container \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": container with ID starting with b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.119696 4935 scope.go:117] "RemoveContainer" containerID="aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.120167 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0"} err="failed to get container status \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": rpc error: code = NotFound desc = could not find container \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": container with ID starting with aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.120190 4935 scope.go:117] "RemoveContainer" containerID="9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.120465 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e"} err="failed to get container status \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": rpc error: code = NotFound desc = could not find container \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": container with ID starting with 9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.120483 4935 scope.go:117] "RemoveContainer" containerID="f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.120780 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50"} err="failed to get container status \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": rpc error: code = NotFound desc = could not find container \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": container with ID starting with f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.120799 4935 scope.go:117] "RemoveContainer" containerID="f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.121146 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c"} err="failed to get container status \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": rpc error: code = NotFound desc = could not find container \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": container with ID starting with f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.121164 4935 scope.go:117] "RemoveContainer" containerID="80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.121566 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa"} err="failed to get container status \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": rpc error: code = NotFound desc = could not find container \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": container with ID starting with 80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.121586 4935 scope.go:117] "RemoveContainer" containerID="3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.122093 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc"} err="failed to get container status \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": rpc error: code = NotFound desc = could not find container \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": container with ID starting with 3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.122115 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.122693 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b"} err="failed to get container status \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": rpc error: code = NotFound desc = could not find container \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": container with ID starting with 1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.122863 4935 scope.go:117] "RemoveContainer" containerID="3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.123439 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5"} err="failed to get container status \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": rpc error: code = NotFound desc = could not find container \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": container with ID starting with 3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.123460 4935 scope.go:117] "RemoveContainer" containerID="8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.123793 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b"} err="failed to get container status \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": rpc error: code = NotFound desc = could not find container \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": container with ID starting with 8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.123827 4935 scope.go:117] "RemoveContainer" containerID="b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.124142 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b"} err="failed to get container status \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": rpc error: code = NotFound desc = could not find container \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": container with ID starting with b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.124176 4935 scope.go:117] "RemoveContainer" containerID="aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.124466 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0"} err="failed to get container status \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": rpc error: code = NotFound desc = could not find container \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": container with ID starting with aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.124503 4935 scope.go:117] "RemoveContainer" containerID="9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.124958 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e"} err="failed to get container status \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": rpc error: code = NotFound desc = could not find container \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": container with ID starting with 9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.124997 4935 scope.go:117] "RemoveContainer" containerID="f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.125375 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50"} err="failed to get container status \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": rpc error: code = NotFound desc = could not find container \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": container with ID starting with f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.125404 4935 scope.go:117] "RemoveContainer" containerID="f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.125683 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c"} err="failed to get container status \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": rpc error: code = NotFound desc = could not find container \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": container with ID starting with f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.125702 4935 scope.go:117] "RemoveContainer" containerID="80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.125934 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa"} err="failed to get container status \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": rpc error: code = NotFound desc = could not find container \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": container with ID starting with 80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.125955 4935 scope.go:117] "RemoveContainer" containerID="3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.126326 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc"} err="failed to get container status \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": rpc error: code = NotFound desc = could not find container \"3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc\": container with ID starting with 3517a506202eaae70c27bab4286f09705d5800b8e53b3ba059d07af56c8f13bc not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.126354 4935 scope.go:117] "RemoveContainer" containerID="1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.126590 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b"} err="failed to get container status \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": rpc error: code = NotFound desc = could not find container \"1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b\": container with ID starting with 1d690f8af965a8025dc04af9504c9d6db18368dfecd1cef95f7e171daada4b1b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.126626 4935 scope.go:117] "RemoveContainer" containerID="3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.127039 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5"} err="failed to get container status \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": rpc error: code = NotFound desc = could not find container \"3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5\": container with ID starting with 3d5af2906a132466ea2b93b44cb50b955b9da8bad6a4e936ee9306167ce3d6a5 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.127068 4935 scope.go:117] "RemoveContainer" containerID="8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.127322 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b"} err="failed to get container status \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": rpc error: code = NotFound desc = could not find container \"8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b\": container with ID starting with 8de802c7a75e10ae1791cec4e7bb056d2fd6b5fa9cc9a14dd6fa78861c4b1d7b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.127346 4935 scope.go:117] "RemoveContainer" containerID="b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.127596 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b"} err="failed to get container status \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": rpc error: code = NotFound desc = could not find container \"b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b\": container with ID starting with b153401feee329881dadb9c90480e2c94f7549ce6f26fd14dd11f9636593994b not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.127617 4935 scope.go:117] "RemoveContainer" containerID="aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.127987 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0"} err="failed to get container status \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": rpc error: code = NotFound desc = could not find container \"aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0\": container with ID starting with aed59a83b4b9d9bfd09c0f9b45eba73f4e6b12436ef2d83f70286deff84d76c0 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.128009 4935 scope.go:117] "RemoveContainer" containerID="9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.128305 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e"} err="failed to get container status \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": rpc error: code = NotFound desc = could not find container \"9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e\": container with ID starting with 9b927148f83ed88e92447b1f482d6fbc5ca1445840792b582c49e16e3892872e not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.128333 4935 scope.go:117] "RemoveContainer" containerID="f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.128603 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50"} err="failed to get container status \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": rpc error: code = NotFound desc = could not find container \"f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50\": container with ID starting with f2753c307e9fd9acd2b7c8fce56fbc6b430d628d2296ed099beb5ccbd2833c50 not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.128641 4935 scope.go:117] "RemoveContainer" containerID="f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.128873 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c"} err="failed to get container status \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": rpc error: code = NotFound desc = could not find container \"f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c\": container with ID starting with f370e9f2fef88e7b35170a8886975bf4e9accccce7916ebd4872550cab4e365c not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.128916 4935 scope.go:117] "RemoveContainer" containerID="80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.129203 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa"} err="failed to get container status \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": rpc error: code = NotFound desc = could not find container \"80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa\": container with ID starting with 80bd37651a2b103c14f9d570673878760e0a5bd446dcc5273e1ca0151d3f6aaa not found: ID does not exist" Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.893929 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"c62a0bbb42bc8b90694cf8a1d933f0a5ed8a184bc332834f9cef0c8b3ee7db96"} Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.894466 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"c779819cd2fb88dc5692506eb8d0850781aabd0c5493f01188cb0c5fd54dab39"} Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.894479 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"c8e1a298ab48a8609c170917dcfdbc73df74ce59d1d8c5763db4ffee289ffa36"} Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.894496 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"f821216bac3f67dc300d6a018f1b857874285e16b842588596fe29a6909f1190"} Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.894509 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"e2c8c7fdc2090b69cd291d1d7ec42b28cf07946e7f044d1a2a94ac9aedaf8952"} Oct 05 07:05:39 crc kubenswrapper[4935]: I1005 07:05:39.894520 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"955479b70b0d6a52987fe907bddf4085ade2ba4254044b0b82f958f69ac4059b"} Oct 05 07:05:40 crc kubenswrapper[4935]: I1005 07:05:40.787120 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="019e7dde-1935-4b4c-8312-99f52b6d22fc" path="/var/lib/kubelet/pods/019e7dde-1935-4b4c-8312-99f52b6d22fc/volumes" Oct 05 07:05:41 crc kubenswrapper[4935]: I1005 07:05:41.913577 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"3c27927e46056ee98647182229c262176c900821ca19ebffde8d8087f8fdaffb"} Oct 05 07:05:44 crc kubenswrapper[4935]: I1005 07:05:44.941112 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" event={"ID":"52cb3655-07ef-4ad8-8828-53ff2491c68c","Type":"ContainerStarted","Data":"10c35dc776ee2a75acfe311c0da6e6f6367737f5272316dd58aae8957e35f16c"} Oct 05 07:05:44 crc kubenswrapper[4935]: I1005 07:05:44.941708 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:44 crc kubenswrapper[4935]: I1005 07:05:44.941723 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:44 crc kubenswrapper[4935]: I1005 07:05:44.941735 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:44 crc kubenswrapper[4935]: I1005 07:05:44.975081 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:44 crc kubenswrapper[4935]: I1005 07:05:44.983656 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" podStartSLOduration=6.983636647 podStartE2EDuration="6.983636647s" podCreationTimestamp="2025-10-05 07:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:05:44.977176746 +0000 UTC m=+778.859803256" watchObservedRunningTime="2025-10-05 07:05:44.983636647 +0000 UTC m=+778.866263117" Oct 05 07:05:44 crc kubenswrapper[4935]: I1005 07:05:44.993939 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.273782 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-j2hvq"] Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.274663 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.276430 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.276779 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.276819 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.276979 4935 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-qpj25" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.286082 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-j2hvq"] Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.433483 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a62d002c-6756-491b-8935-02235d870828-node-mnt\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.433532 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a62d002c-6756-491b-8935-02235d870828-crc-storage\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.433557 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twvc7\" (UniqueName: \"kubernetes.io/projected/a62d002c-6756-491b-8935-02235d870828-kube-api-access-twvc7\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.535094 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a62d002c-6756-491b-8935-02235d870828-node-mnt\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.535170 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a62d002c-6756-491b-8935-02235d870828-crc-storage\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.535225 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twvc7\" (UniqueName: \"kubernetes.io/projected/a62d002c-6756-491b-8935-02235d870828-kube-api-access-twvc7\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.535465 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a62d002c-6756-491b-8935-02235d870828-node-mnt\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.536027 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a62d002c-6756-491b-8935-02235d870828-crc-storage\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.554685 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twvc7\" (UniqueName: \"kubernetes.io/projected/a62d002c-6756-491b-8935-02235d870828-kube-api-access-twvc7\") pod \"crc-storage-crc-j2hvq\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.592597 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:47 crc kubenswrapper[4935]: I1005 07:05:47.983990 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-j2hvq"] Oct 05 07:05:47 crc kubenswrapper[4935]: W1005 07:05:47.999197 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda62d002c_6756_491b_8935_02235d870828.slice/crio-cdfffadccd52199648bcb026f6883985fbcb4c0ded8becff2b91bfa9b93169ad WatchSource:0}: Error finding container cdfffadccd52199648bcb026f6883985fbcb4c0ded8becff2b91bfa9b93169ad: Status 404 returned error can't find the container with id cdfffadccd52199648bcb026f6883985fbcb4c0ded8becff2b91bfa9b93169ad Oct 05 07:05:48 crc kubenswrapper[4935]: I1005 07:05:48.002660 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:05:48 crc kubenswrapper[4935]: I1005 07:05:48.963251 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-j2hvq" event={"ID":"a62d002c-6756-491b-8935-02235d870828","Type":"ContainerStarted","Data":"cdfffadccd52199648bcb026f6883985fbcb4c0ded8becff2b91bfa9b93169ad"} Oct 05 07:05:49 crc kubenswrapper[4935]: I1005 07:05:49.970161 4935 generic.go:334] "Generic (PLEG): container finished" podID="a62d002c-6756-491b-8935-02235d870828" containerID="d7ffcb2306d39f5d2b1224f80f8160485b1405feabc0a2bd0fb06694d9f3af3a" exitCode=0 Oct 05 07:05:49 crc kubenswrapper[4935]: I1005 07:05:49.970267 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-j2hvq" event={"ID":"a62d002c-6756-491b-8935-02235d870828","Type":"ContainerDied","Data":"d7ffcb2306d39f5d2b1224f80f8160485b1405feabc0a2bd0fb06694d9f3af3a"} Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.355793 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.495815 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a62d002c-6756-491b-8935-02235d870828-node-mnt\") pod \"a62d002c-6756-491b-8935-02235d870828\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.495930 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twvc7\" (UniqueName: \"kubernetes.io/projected/a62d002c-6756-491b-8935-02235d870828-kube-api-access-twvc7\") pod \"a62d002c-6756-491b-8935-02235d870828\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.495965 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a62d002c-6756-491b-8935-02235d870828-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "a62d002c-6756-491b-8935-02235d870828" (UID: "a62d002c-6756-491b-8935-02235d870828"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.496015 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a62d002c-6756-491b-8935-02235d870828-crc-storage\") pod \"a62d002c-6756-491b-8935-02235d870828\" (UID: \"a62d002c-6756-491b-8935-02235d870828\") " Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.496340 4935 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a62d002c-6756-491b-8935-02235d870828-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.502591 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a62d002c-6756-491b-8935-02235d870828-kube-api-access-twvc7" (OuterVolumeSpecName: "kube-api-access-twvc7") pod "a62d002c-6756-491b-8935-02235d870828" (UID: "a62d002c-6756-491b-8935-02235d870828"). InnerVolumeSpecName "kube-api-access-twvc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.512963 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a62d002c-6756-491b-8935-02235d870828-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "a62d002c-6756-491b-8935-02235d870828" (UID: "a62d002c-6756-491b-8935-02235d870828"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.597459 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twvc7\" (UniqueName: \"kubernetes.io/projected/a62d002c-6756-491b-8935-02235d870828-kube-api-access-twvc7\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.597519 4935 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a62d002c-6756-491b-8935-02235d870828-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.886551 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9dqn5"] Oct 05 07:05:51 crc kubenswrapper[4935]: E1005 07:05:51.887147 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a62d002c-6756-491b-8935-02235d870828" containerName="storage" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.887167 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a62d002c-6756-491b-8935-02235d870828" containerName="storage" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.887310 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a62d002c-6756-491b-8935-02235d870828" containerName="storage" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.888963 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.893211 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9dqn5"] Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.981238 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-j2hvq" event={"ID":"a62d002c-6756-491b-8935-02235d870828","Type":"ContainerDied","Data":"cdfffadccd52199648bcb026f6883985fbcb4c0ded8becff2b91bfa9b93169ad"} Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.981274 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdfffadccd52199648bcb026f6883985fbcb4c0ded8becff2b91bfa9b93169ad" Oct 05 07:05:51 crc kubenswrapper[4935]: I1005 07:05:51.981292 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j2hvq" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.002866 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-utilities\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.002990 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbcrd\" (UniqueName: \"kubernetes.io/projected/f8b7483f-b718-42d8-95e6-950acfab4d30-kube-api-access-vbcrd\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.003098 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-catalog-content\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.104446 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-catalog-content\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.104812 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-utilities\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.104869 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-catalog-content\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.104946 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbcrd\" (UniqueName: \"kubernetes.io/projected/f8b7483f-b718-42d8-95e6-950acfab4d30-kube-api-access-vbcrd\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.105195 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-utilities\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.120880 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbcrd\" (UniqueName: \"kubernetes.io/projected/f8b7483f-b718-42d8-95e6-950acfab4d30-kube-api-access-vbcrd\") pod \"community-operators-9dqn5\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.213867 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.643915 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9dqn5"] Oct 05 07:05:52 crc kubenswrapper[4935]: I1005 07:05:52.987027 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9dqn5" event={"ID":"f8b7483f-b718-42d8-95e6-950acfab4d30","Type":"ContainerStarted","Data":"1caacf035ee036139e488ab4e62b35c7e7c09c9182beac16b533dc1d7da3a15b"} Oct 05 07:05:53 crc kubenswrapper[4935]: I1005 07:05:53.993723 4935 generic.go:334] "Generic (PLEG): container finished" podID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerID="8de0256dd007a5a553144b46aad348d4313786ac0936ce0e5e7a3bf1b3212b3c" exitCode=0 Oct 05 07:05:53 crc kubenswrapper[4935]: I1005 07:05:53.993778 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9dqn5" event={"ID":"f8b7483f-b718-42d8-95e6-950acfab4d30","Type":"ContainerDied","Data":"8de0256dd007a5a553144b46aad348d4313786ac0936ce0e5e7a3bf1b3212b3c"} Oct 05 07:05:55 crc kubenswrapper[4935]: I1005 07:05:55.004770 4935 generic.go:334] "Generic (PLEG): container finished" podID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerID="1b99a395eca4a926f0911f71ad49e3c7b4f9a7ca96d1fa3f05b7b1557b0ef1f9" exitCode=0 Oct 05 07:05:55 crc kubenswrapper[4935]: I1005 07:05:55.004827 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9dqn5" event={"ID":"f8b7483f-b718-42d8-95e6-950acfab4d30","Type":"ContainerDied","Data":"1b99a395eca4a926f0911f71ad49e3c7b4f9a7ca96d1fa3f05b7b1557b0ef1f9"} Oct 05 07:05:56 crc kubenswrapper[4935]: I1005 07:05:56.013800 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9dqn5" event={"ID":"f8b7483f-b718-42d8-95e6-950acfab4d30","Type":"ContainerStarted","Data":"dfe62c90fa335a0ed334f9839281e87b352861197b736d726472fd5a4328c441"} Oct 05 07:05:56 crc kubenswrapper[4935]: I1005 07:05:56.036385 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9dqn5" podStartSLOduration=3.59409798 podStartE2EDuration="5.036366978s" podCreationTimestamp="2025-10-05 07:05:51 +0000 UTC" firstStartedPulling="2025-10-05 07:05:53.995328016 +0000 UTC m=+787.877954476" lastFinishedPulling="2025-10-05 07:05:55.437597014 +0000 UTC m=+789.320223474" observedRunningTime="2025-10-05 07:05:56.031723725 +0000 UTC m=+789.914350205" watchObservedRunningTime="2025-10-05 07:05:56.036366978 +0000 UTC m=+789.918993438" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.502224 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz"] Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.503338 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.505142 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.520983 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz"] Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.599390 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.599514 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.599571 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tf7x\" (UniqueName: \"kubernetes.io/projected/3553056f-ef6b-407c-8519-255e0f59496c-kube-api-access-7tf7x\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.701111 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.701157 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tf7x\" (UniqueName: \"kubernetes.io/projected/3553056f-ef6b-407c-8519-255e0f59496c-kube-api-access-7tf7x\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.701232 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.701749 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.701842 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.740953 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tf7x\" (UniqueName: \"kubernetes.io/projected/3553056f-ef6b-407c-8519-255e0f59496c-kube-api-access-7tf7x\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:58 crc kubenswrapper[4935]: I1005 07:05:58.818390 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:05:59 crc kubenswrapper[4935]: I1005 07:05:59.184649 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz"] Oct 05 07:05:59 crc kubenswrapper[4935]: W1005 07:05:59.195650 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3553056f_ef6b_407c_8519_255e0f59496c.slice/crio-6f530badf95d7093689a03a3f08c727cb0164bf4962ae83e71692c97e4043c78 WatchSource:0}: Error finding container 6f530badf95d7093689a03a3f08c727cb0164bf4962ae83e71692c97e4043c78: Status 404 returned error can't find the container with id 6f530badf95d7093689a03a3f08c727cb0164bf4962ae83e71692c97e4043c78 Oct 05 07:06:00 crc kubenswrapper[4935]: I1005 07:06:00.034513 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" event={"ID":"3553056f-ef6b-407c-8519-255e0f59496c","Type":"ContainerStarted","Data":"9dcf80fa24a8b52ab1506780f9dad5beb51ccee338f111962f02b4093f61941d"} Oct 05 07:06:00 crc kubenswrapper[4935]: I1005 07:06:00.034586 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" event={"ID":"3553056f-ef6b-407c-8519-255e0f59496c","Type":"ContainerStarted","Data":"6f530badf95d7093689a03a3f08c727cb0164bf4962ae83e71692c97e4043c78"} Oct 05 07:06:00 crc kubenswrapper[4935]: I1005 07:06:00.871076 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6qnhq"] Oct 05 07:06:00 crc kubenswrapper[4935]: I1005 07:06:00.873650 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:00 crc kubenswrapper[4935]: I1005 07:06:00.879941 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6qnhq"] Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.032468 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m58z2\" (UniqueName: \"kubernetes.io/projected/089298bd-a478-4011-9320-1f78ed045035-kube-api-access-m58z2\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.032572 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-catalog-content\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.032693 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-utilities\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.040264 4935 generic.go:334] "Generic (PLEG): container finished" podID="3553056f-ef6b-407c-8519-255e0f59496c" containerID="9dcf80fa24a8b52ab1506780f9dad5beb51ccee338f111962f02b4093f61941d" exitCode=0 Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.040314 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" event={"ID":"3553056f-ef6b-407c-8519-255e0f59496c","Type":"ContainerDied","Data":"9dcf80fa24a8b52ab1506780f9dad5beb51ccee338f111962f02b4093f61941d"} Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.133443 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-catalog-content\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.133497 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-utilities\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.133580 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m58z2\" (UniqueName: \"kubernetes.io/projected/089298bd-a478-4011-9320-1f78ed045035-kube-api-access-m58z2\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.134146 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-catalog-content\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.134318 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-utilities\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.155336 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m58z2\" (UniqueName: \"kubernetes.io/projected/089298bd-a478-4011-9320-1f78ed045035-kube-api-access-m58z2\") pod \"redhat-operators-6qnhq\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.198905 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:01 crc kubenswrapper[4935]: I1005 07:06:01.624470 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6qnhq"] Oct 05 07:06:01 crc kubenswrapper[4935]: W1005 07:06:01.631525 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod089298bd_a478_4011_9320_1f78ed045035.slice/crio-f788774d33f3dd84d6e0a7f764d30924e94365527f97e7686c25f0619815737b WatchSource:0}: Error finding container f788774d33f3dd84d6e0a7f764d30924e94365527f97e7686c25f0619815737b: Status 404 returned error can't find the container with id f788774d33f3dd84d6e0a7f764d30924e94365527f97e7686c25f0619815737b Oct 05 07:06:02 crc kubenswrapper[4935]: I1005 07:06:02.046884 4935 generic.go:334] "Generic (PLEG): container finished" podID="089298bd-a478-4011-9320-1f78ed045035" containerID="269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335" exitCode=0 Oct 05 07:06:02 crc kubenswrapper[4935]: I1005 07:06:02.047537 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qnhq" event={"ID":"089298bd-a478-4011-9320-1f78ed045035","Type":"ContainerDied","Data":"269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335"} Oct 05 07:06:02 crc kubenswrapper[4935]: I1005 07:06:02.047672 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qnhq" event={"ID":"089298bd-a478-4011-9320-1f78ed045035","Type":"ContainerStarted","Data":"f788774d33f3dd84d6e0a7f764d30924e94365527f97e7686c25f0619815737b"} Oct 05 07:06:02 crc kubenswrapper[4935]: I1005 07:06:02.214734 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:06:02 crc kubenswrapper[4935]: I1005 07:06:02.214790 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:06:02 crc kubenswrapper[4935]: I1005 07:06:02.259366 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:06:03 crc kubenswrapper[4935]: I1005 07:06:03.130311 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:06:04 crc kubenswrapper[4935]: I1005 07:06:04.068569 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" event={"ID":"3553056f-ef6b-407c-8519-255e0f59496c","Type":"ContainerStarted","Data":"eb3416ccfe9251f527b06554b9fad0e440f896bf3564c8e3d02ca77de4864262"} Oct 05 07:06:05 crc kubenswrapper[4935]: I1005 07:06:05.077686 4935 generic.go:334] "Generic (PLEG): container finished" podID="3553056f-ef6b-407c-8519-255e0f59496c" containerID="eb3416ccfe9251f527b06554b9fad0e440f896bf3564c8e3d02ca77de4864262" exitCode=0 Oct 05 07:06:05 crc kubenswrapper[4935]: I1005 07:06:05.077808 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" event={"ID":"3553056f-ef6b-407c-8519-255e0f59496c","Type":"ContainerDied","Data":"eb3416ccfe9251f527b06554b9fad0e440f896bf3564c8e3d02ca77de4864262"} Oct 05 07:06:05 crc kubenswrapper[4935]: I1005 07:06:05.083256 4935 generic.go:334] "Generic (PLEG): container finished" podID="089298bd-a478-4011-9320-1f78ed045035" containerID="d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6" exitCode=0 Oct 05 07:06:05 crc kubenswrapper[4935]: I1005 07:06:05.083309 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qnhq" event={"ID":"089298bd-a478-4011-9320-1f78ed045035","Type":"ContainerDied","Data":"d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6"} Oct 05 07:06:05 crc kubenswrapper[4935]: I1005 07:06:05.857721 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9dqn5"] Oct 05 07:06:05 crc kubenswrapper[4935]: I1005 07:06:05.858217 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9dqn5" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerName="registry-server" containerID="cri-o://dfe62c90fa335a0ed334f9839281e87b352861197b736d726472fd5a4328c441" gracePeriod=2 Oct 05 07:06:06 crc kubenswrapper[4935]: I1005 07:06:06.092485 4935 generic.go:334] "Generic (PLEG): container finished" podID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerID="dfe62c90fa335a0ed334f9839281e87b352861197b736d726472fd5a4328c441" exitCode=0 Oct 05 07:06:06 crc kubenswrapper[4935]: I1005 07:06:06.092537 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9dqn5" event={"ID":"f8b7483f-b718-42d8-95e6-950acfab4d30","Type":"ContainerDied","Data":"dfe62c90fa335a0ed334f9839281e87b352861197b736d726472fd5a4328c441"} Oct 05 07:06:06 crc kubenswrapper[4935]: I1005 07:06:06.096032 4935 generic.go:334] "Generic (PLEG): container finished" podID="3553056f-ef6b-407c-8519-255e0f59496c" containerID="763f52b2005d596453353988d97c20d170747572865648dfd91069574c691748" exitCode=0 Oct 05 07:06:06 crc kubenswrapper[4935]: I1005 07:06:06.096090 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" event={"ID":"3553056f-ef6b-407c-8519-255e0f59496c","Type":"ContainerDied","Data":"763f52b2005d596453353988d97c20d170747572865648dfd91069574c691748"} Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.103729 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9dqn5" event={"ID":"f8b7483f-b718-42d8-95e6-950acfab4d30","Type":"ContainerDied","Data":"1caacf035ee036139e488ab4e62b35c7e7c09c9182beac16b533dc1d7da3a15b"} Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.104128 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1caacf035ee036139e488ab4e62b35c7e7c09c9182beac16b533dc1d7da3a15b" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.106862 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qnhq" event={"ID":"089298bd-a478-4011-9320-1f78ed045035","Type":"ContainerStarted","Data":"0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8"} Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.124228 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6qnhq" podStartSLOduration=2.2561870219999998 podStartE2EDuration="7.12421319s" podCreationTimestamp="2025-10-05 07:06:00 +0000 UTC" firstStartedPulling="2025-10-05 07:06:02.048697971 +0000 UTC m=+795.931324441" lastFinishedPulling="2025-10-05 07:06:06.916724159 +0000 UTC m=+800.799350609" observedRunningTime="2025-10-05 07:06:07.123410209 +0000 UTC m=+801.006036669" watchObservedRunningTime="2025-10-05 07:06:07.12421319 +0000 UTC m=+801.006839650" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.139197 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.308475 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-catalog-content\") pod \"f8b7483f-b718-42d8-95e6-950acfab4d30\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.308558 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-utilities\") pod \"f8b7483f-b718-42d8-95e6-950acfab4d30\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.308625 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbcrd\" (UniqueName: \"kubernetes.io/projected/f8b7483f-b718-42d8-95e6-950acfab4d30-kube-api-access-vbcrd\") pod \"f8b7483f-b718-42d8-95e6-950acfab4d30\" (UID: \"f8b7483f-b718-42d8-95e6-950acfab4d30\") " Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.309453 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-utilities" (OuterVolumeSpecName: "utilities") pod "f8b7483f-b718-42d8-95e6-950acfab4d30" (UID: "f8b7483f-b718-42d8-95e6-950acfab4d30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.314092 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8b7483f-b718-42d8-95e6-950acfab4d30-kube-api-access-vbcrd" (OuterVolumeSpecName: "kube-api-access-vbcrd") pod "f8b7483f-b718-42d8-95e6-950acfab4d30" (UID: "f8b7483f-b718-42d8-95e6-950acfab4d30"). InnerVolumeSpecName "kube-api-access-vbcrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.351509 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.357593 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8b7483f-b718-42d8-95e6-950acfab4d30" (UID: "f8b7483f-b718-42d8-95e6-950acfab4d30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.410023 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.410055 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b7483f-b718-42d8-95e6-950acfab4d30-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.410067 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbcrd\" (UniqueName: \"kubernetes.io/projected/f8b7483f-b718-42d8-95e6-950acfab4d30-kube-api-access-vbcrd\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.510658 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-util\") pod \"3553056f-ef6b-407c-8519-255e0f59496c\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.510975 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-bundle\") pod \"3553056f-ef6b-407c-8519-255e0f59496c\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.511022 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tf7x\" (UniqueName: \"kubernetes.io/projected/3553056f-ef6b-407c-8519-255e0f59496c-kube-api-access-7tf7x\") pod \"3553056f-ef6b-407c-8519-255e0f59496c\" (UID: \"3553056f-ef6b-407c-8519-255e0f59496c\") " Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.511424 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-bundle" (OuterVolumeSpecName: "bundle") pod "3553056f-ef6b-407c-8519-255e0f59496c" (UID: "3553056f-ef6b-407c-8519-255e0f59496c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.514149 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3553056f-ef6b-407c-8519-255e0f59496c-kube-api-access-7tf7x" (OuterVolumeSpecName: "kube-api-access-7tf7x") pod "3553056f-ef6b-407c-8519-255e0f59496c" (UID: "3553056f-ef6b-407c-8519-255e0f59496c"). InnerVolumeSpecName "kube-api-access-7tf7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.520084 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-util" (OuterVolumeSpecName: "util") pod "3553056f-ef6b-407c-8519-255e0f59496c" (UID: "3553056f-ef6b-407c-8519-255e0f59496c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.611740 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tf7x\" (UniqueName: \"kubernetes.io/projected/3553056f-ef6b-407c-8519-255e0f59496c-kube-api-access-7tf7x\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.611777 4935 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-util\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:07 crc kubenswrapper[4935]: I1005 07:06:07.611786 4935 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3553056f-ef6b-407c-8519-255e0f59496c-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:08 crc kubenswrapper[4935]: I1005 07:06:08.117834 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" Oct 05 07:06:08 crc kubenswrapper[4935]: I1005 07:06:08.117822 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz" event={"ID":"3553056f-ef6b-407c-8519-255e0f59496c","Type":"ContainerDied","Data":"6f530badf95d7093689a03a3f08c727cb0164bf4962ae83e71692c97e4043c78"} Oct 05 07:06:08 crc kubenswrapper[4935]: I1005 07:06:08.117882 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f530badf95d7093689a03a3f08c727cb0164bf4962ae83e71692c97e4043c78" Oct 05 07:06:08 crc kubenswrapper[4935]: I1005 07:06:08.117882 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9dqn5" Oct 05 07:06:08 crc kubenswrapper[4935]: I1005 07:06:08.150342 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9dqn5"] Oct 05 07:06:08 crc kubenswrapper[4935]: I1005 07:06:08.156119 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9dqn5"] Oct 05 07:06:08 crc kubenswrapper[4935]: I1005 07:06:08.421268 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pbfk4" Oct 05 07:06:08 crc kubenswrapper[4935]: I1005 07:06:08.786181 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" path="/var/lib/kubelet/pods/f8b7483f-b718-42d8-95e6-950acfab4d30/volumes" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476126 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-x2q48"] Oct 05 07:06:09 crc kubenswrapper[4935]: E1005 07:06:09.476377 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3553056f-ef6b-407c-8519-255e0f59496c" containerName="extract" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476392 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3553056f-ef6b-407c-8519-255e0f59496c" containerName="extract" Oct 05 07:06:09 crc kubenswrapper[4935]: E1005 07:06:09.476404 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3553056f-ef6b-407c-8519-255e0f59496c" containerName="pull" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476412 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3553056f-ef6b-407c-8519-255e0f59496c" containerName="pull" Oct 05 07:06:09 crc kubenswrapper[4935]: E1005 07:06:09.476424 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3553056f-ef6b-407c-8519-255e0f59496c" containerName="util" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476431 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3553056f-ef6b-407c-8519-255e0f59496c" containerName="util" Oct 05 07:06:09 crc kubenswrapper[4935]: E1005 07:06:09.476441 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerName="registry-server" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476449 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerName="registry-server" Oct 05 07:06:09 crc kubenswrapper[4935]: E1005 07:06:09.476463 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerName="extract-utilities" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476470 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerName="extract-utilities" Oct 05 07:06:09 crc kubenswrapper[4935]: E1005 07:06:09.476484 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerName="extract-content" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476492 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerName="extract-content" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476603 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8b7483f-b718-42d8-95e6-950acfab4d30" containerName="registry-server" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.476620 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="3553056f-ef6b-407c-8519-255e0f59496c" containerName="extract" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.477107 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x2q48" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.479943 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-ttf9f" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.480536 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.480730 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.487230 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-x2q48"] Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.534819 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvf9z\" (UniqueName: \"kubernetes.io/projected/7f754f36-a0e0-4a97-b03d-f8d5d69f4b00-kube-api-access-lvf9z\") pod \"nmstate-operator-858ddd8f98-x2q48\" (UID: \"7f754f36-a0e0-4a97-b03d-f8d5d69f4b00\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-x2q48" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.635684 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvf9z\" (UniqueName: \"kubernetes.io/projected/7f754f36-a0e0-4a97-b03d-f8d5d69f4b00-kube-api-access-lvf9z\") pod \"nmstate-operator-858ddd8f98-x2q48\" (UID: \"7f754f36-a0e0-4a97-b03d-f8d5d69f4b00\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-x2q48" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.659055 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvf9z\" (UniqueName: \"kubernetes.io/projected/7f754f36-a0e0-4a97-b03d-f8d5d69f4b00-kube-api-access-lvf9z\") pod \"nmstate-operator-858ddd8f98-x2q48\" (UID: \"7f754f36-a0e0-4a97-b03d-f8d5d69f4b00\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-x2q48" Oct 05 07:06:09 crc kubenswrapper[4935]: I1005 07:06:09.792231 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x2q48" Oct 05 07:06:10 crc kubenswrapper[4935]: I1005 07:06:10.194548 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-x2q48"] Oct 05 07:06:11 crc kubenswrapper[4935]: I1005 07:06:11.130840 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x2q48" event={"ID":"7f754f36-a0e0-4a97-b03d-f8d5d69f4b00","Type":"ContainerStarted","Data":"ccb092069a228f48868ea87584cd1f3934245de67f31e4a24a251747f4678583"} Oct 05 07:06:11 crc kubenswrapper[4935]: I1005 07:06:11.199445 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:11 crc kubenswrapper[4935]: I1005 07:06:11.199496 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:12 crc kubenswrapper[4935]: I1005 07:06:12.238483 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6qnhq" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="registry-server" probeResult="failure" output=< Oct 05 07:06:12 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 07:06:12 crc kubenswrapper[4935]: > Oct 05 07:06:13 crc kubenswrapper[4935]: I1005 07:06:13.144981 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x2q48" event={"ID":"7f754f36-a0e0-4a97-b03d-f8d5d69f4b00","Type":"ContainerStarted","Data":"27a121ccf7b25b5fcb8b6078c9d5a5d1ddd7b97a779f0dd36a523cf2dd35b9ab"} Oct 05 07:06:13 crc kubenswrapper[4935]: I1005 07:06:13.175243 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x2q48" podStartSLOduration=2.101790509 podStartE2EDuration="4.175218431s" podCreationTimestamp="2025-10-05 07:06:09 +0000 UTC" firstStartedPulling="2025-10-05 07:06:10.205619455 +0000 UTC m=+804.088245905" lastFinishedPulling="2025-10-05 07:06:12.279047367 +0000 UTC m=+806.161673827" observedRunningTime="2025-10-05 07:06:13.169201751 +0000 UTC m=+807.051828211" watchObservedRunningTime="2025-10-05 07:06:13.175218431 +0000 UTC m=+807.057844931" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.731635 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9"] Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.732910 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" Oct 05 07:06:18 crc kubenswrapper[4935]: W1005 07:06:18.735206 4935 reflector.go:561] object-"openshift-nmstate"/"nmstate-handler-dockercfg-x8tb4": failed to list *v1.Secret: secrets "nmstate-handler-dockercfg-x8tb4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Oct 05 07:06:18 crc kubenswrapper[4935]: E1005 07:06:18.735256 4935 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"nmstate-handler-dockercfg-x8tb4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nmstate-handler-dockercfg-x8tb4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.743177 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9"] Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.758231 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-n8448"] Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.759030 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.764110 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.773204 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-qxgvc"] Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.774240 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.819864 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-n8448"] Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.847762 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6xw4\" (UniqueName: \"kubernetes.io/projected/2a702a6b-c2f1-46ae-b6da-a99ff519a83f-kube-api-access-v6xw4\") pod \"nmstate-metrics-fdff9cb8d-blpj9\" (UID: \"2a702a6b-c2f1-46ae-b6da-a99ff519a83f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.876854 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm"] Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.877575 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.879672 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.880278 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.882659 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-2s7hj" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.885524 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm"] Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.949525 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-ovs-socket\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.949560 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgtfg\" (UniqueName: \"kubernetes.io/projected/66a68296-c0ef-463c-8d6f-e6a970833180-kube-api-access-xgtfg\") pod \"nmstate-webhook-6cdbc54649-n8448\" (UID: \"66a68296-c0ef-463c-8d6f-e6a970833180\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.949585 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/66a68296-c0ef-463c-8d6f-e6a970833180-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-n8448\" (UID: \"66a68296-c0ef-463c-8d6f-e6a970833180\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.949829 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6xw4\" (UniqueName: \"kubernetes.io/projected/2a702a6b-c2f1-46ae-b6da-a99ff519a83f-kube-api-access-v6xw4\") pod \"nmstate-metrics-fdff9cb8d-blpj9\" (UID: \"2a702a6b-c2f1-46ae-b6da-a99ff519a83f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.949881 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-dbus-socket\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.949926 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-nmstate-lock\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.949950 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b2lc\" (UniqueName: \"kubernetes.io/projected/7d9b8112-d308-439e-838d-410a3bc6bf18-kube-api-access-4b2lc\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:18 crc kubenswrapper[4935]: I1005 07:06:18.969117 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6xw4\" (UniqueName: \"kubernetes.io/projected/2a702a6b-c2f1-46ae-b6da-a99ff519a83f-kube-api-access-v6xw4\") pod \"nmstate-metrics-fdff9cb8d-blpj9\" (UID: \"2a702a6b-c2f1-46ae-b6da-a99ff519a83f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050609 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-ovs-socket\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050657 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgtfg\" (UniqueName: \"kubernetes.io/projected/66a68296-c0ef-463c-8d6f-e6a970833180-kube-api-access-xgtfg\") pod \"nmstate-webhook-6cdbc54649-n8448\" (UID: \"66a68296-c0ef-463c-8d6f-e6a970833180\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050682 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/66a68296-c0ef-463c-8d6f-e6a970833180-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-n8448\" (UID: \"66a68296-c0ef-463c-8d6f-e6a970833180\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050714 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-dbus-socket\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050736 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/406136d1-87c2-4e34-9941-38d84f0aaff4-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050758 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-nmstate-lock\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050773 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b2lc\" (UniqueName: \"kubernetes.io/projected/7d9b8112-d308-439e-838d-410a3bc6bf18-kube-api-access-4b2lc\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050800 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgjhq\" (UniqueName: \"kubernetes.io/projected/406136d1-87c2-4e34-9941-38d84f0aaff4-kube-api-access-fgjhq\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050835 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/406136d1-87c2-4e34-9941-38d84f0aaff4-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.050941 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-nmstate-lock\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.051111 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-dbus-socket\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.051351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7d9b8112-d308-439e-838d-410a3bc6bf18-ovs-socket\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.053850 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/66a68296-c0ef-463c-8d6f-e6a970833180-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-n8448\" (UID: \"66a68296-c0ef-463c-8d6f-e6a970833180\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.059408 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-78c49f5cb6-zvfvk"] Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.060243 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.072600 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgtfg\" (UniqueName: \"kubernetes.io/projected/66a68296-c0ef-463c-8d6f-e6a970833180-kube-api-access-xgtfg\") pod \"nmstate-webhook-6cdbc54649-n8448\" (UID: \"66a68296-c0ef-463c-8d6f-e6a970833180\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.078098 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78c49f5cb6-zvfvk"] Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.083467 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b2lc\" (UniqueName: \"kubernetes.io/projected/7d9b8112-d308-439e-838d-410a3bc6bf18-kube-api-access-4b2lc\") pod \"nmstate-handler-qxgvc\" (UID: \"7d9b8112-d308-439e-838d-410a3bc6bf18\") " pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151516 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-service-ca\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151565 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/406136d1-87c2-4e34-9941-38d84f0aaff4-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151589 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq7l2\" (UniqueName: \"kubernetes.io/projected/61bd0e6f-248c-4140-99ae-825b3a143133-kube-api-access-mq7l2\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151616 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgjhq\" (UniqueName: \"kubernetes.io/projected/406136d1-87c2-4e34-9941-38d84f0aaff4-kube-api-access-fgjhq\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151644 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-trusted-ca-bundle\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151661 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-oauth-serving-cert\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151677 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/406136d1-87c2-4e34-9941-38d84f0aaff4-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151708 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/61bd0e6f-248c-4140-99ae-825b3a143133-console-oauth-config\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151727 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/61bd0e6f-248c-4140-99ae-825b3a143133-console-serving-cert\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.151746 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-console-config\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: E1005 07:06:19.151860 4935 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 05 07:06:19 crc kubenswrapper[4935]: E1005 07:06:19.151947 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/406136d1-87c2-4e34-9941-38d84f0aaff4-plugin-serving-cert podName:406136d1-87c2-4e34-9941-38d84f0aaff4 nodeName:}" failed. No retries permitted until 2025-10-05 07:06:19.651931017 +0000 UTC m=+813.534557477 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/406136d1-87c2-4e34-9941-38d84f0aaff4-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-bl5mm" (UID: "406136d1-87c2-4e34-9941-38d84f0aaff4") : secret "plugin-serving-cert" not found Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.152492 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/406136d1-87c2-4e34-9941-38d84f0aaff4-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.169351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgjhq\" (UniqueName: \"kubernetes.io/projected/406136d1-87c2-4e34-9941-38d84f0aaff4-kube-api-access-fgjhq\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.252746 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-service-ca\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.252811 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq7l2\" (UniqueName: \"kubernetes.io/projected/61bd0e6f-248c-4140-99ae-825b3a143133-kube-api-access-mq7l2\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.252860 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-trusted-ca-bundle\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.252879 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-oauth-serving-cert\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.252952 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/61bd0e6f-248c-4140-99ae-825b3a143133-console-oauth-config\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.252980 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/61bd0e6f-248c-4140-99ae-825b3a143133-console-serving-cert\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.253000 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-console-config\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.253771 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-service-ca\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.254009 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-oauth-serving-cert\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.254300 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-console-config\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.254651 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61bd0e6f-248c-4140-99ae-825b3a143133-trusted-ca-bundle\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.255942 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/61bd0e6f-248c-4140-99ae-825b3a143133-console-oauth-config\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.256421 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/61bd0e6f-248c-4140-99ae-825b3a143133-console-serving-cert\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.267141 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq7l2\" (UniqueName: \"kubernetes.io/projected/61bd0e6f-248c-4140-99ae-825b3a143133-kube-api-access-mq7l2\") pod \"console-78c49f5cb6-zvfvk\" (UID: \"61bd0e6f-248c-4140-99ae-825b3a143133\") " pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.409436 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.662575 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/406136d1-87c2-4e34-9941-38d84f0aaff4-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.667863 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/406136d1-87c2-4e34-9941-38d84f0aaff4-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-bl5mm\" (UID: \"406136d1-87c2-4e34-9941-38d84f0aaff4\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.791065 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.801292 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78c49f5cb6-zvfvk"] Oct 05 07:06:19 crc kubenswrapper[4935]: I1005 07:06:19.993329 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm"] Oct 05 07:06:19 crc kubenswrapper[4935]: W1005 07:06:19.999257 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod406136d1_87c2_4e34_9941_38d84f0aaff4.slice/crio-827a9bf83b455cf0b8aa4ea9ed5b1caa712413a69e0a27ecd1167eea961a20ce WatchSource:0}: Error finding container 827a9bf83b455cf0b8aa4ea9ed5b1caa712413a69e0a27ecd1167eea961a20ce: Status 404 returned error can't find the container with id 827a9bf83b455cf0b8aa4ea9ed5b1caa712413a69e0a27ecd1167eea961a20ce Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.048263 4935 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.048305 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.109069 4935 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-nmstate/nmstate-handler-qxgvc" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.109142 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.200420 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78c49f5cb6-zvfvk" event={"ID":"61bd0e6f-248c-4140-99ae-825b3a143133","Type":"ContainerStarted","Data":"42474171e3e779026573da4aee0a1f55a6dc54b0334b90c2d223bb9d9d05ba5f"} Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.200769 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78c49f5cb6-zvfvk" event={"ID":"61bd0e6f-248c-4140-99ae-825b3a143133","Type":"ContainerStarted","Data":"9380dda635b2b5b13d4a13b14406b0335249eb9bd388293dda737745d04c1a39"} Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.201631 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" event={"ID":"406136d1-87c2-4e34-9941-38d84f0aaff4","Type":"ContainerStarted","Data":"827a9bf83b455cf0b8aa4ea9ed5b1caa712413a69e0a27ecd1167eea961a20ce"} Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.216567 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-78c49f5cb6-zvfvk" podStartSLOduration=1.216547775 podStartE2EDuration="1.216547775s" podCreationTimestamp="2025-10-05 07:06:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:06:20.21488126 +0000 UTC m=+814.097507740" watchObservedRunningTime="2025-10-05 07:06:20.216547775 +0000 UTC m=+814.099174235" Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.225970 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-x8tb4" Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.231909 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.526814 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9"] Oct 05 07:06:20 crc kubenswrapper[4935]: W1005 07:06:20.530008 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a702a6b_c2f1_46ae_b6da_a99ff519a83f.slice/crio-c4bdd2f40d57c2cc26793b86315d5c62ee0c0df1729666774a87b1813da280de WatchSource:0}: Error finding container c4bdd2f40d57c2cc26793b86315d5c62ee0c0df1729666774a87b1813da280de: Status 404 returned error can't find the container with id c4bdd2f40d57c2cc26793b86315d5c62ee0c0df1729666774a87b1813da280de Oct 05 07:06:20 crc kubenswrapper[4935]: I1005 07:06:20.623972 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-n8448"] Oct 05 07:06:20 crc kubenswrapper[4935]: W1005 07:06:20.632921 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66a68296_c0ef_463c_8d6f_e6a970833180.slice/crio-ff56aecec95681d64e15a837bcf70d0b408ffb25ed36cc7e78952dcd15565200 WatchSource:0}: Error finding container ff56aecec95681d64e15a837bcf70d0b408ffb25ed36cc7e78952dcd15565200: Status 404 returned error can't find the container with id ff56aecec95681d64e15a837bcf70d0b408ffb25ed36cc7e78952dcd15565200 Oct 05 07:06:21 crc kubenswrapper[4935]: I1005 07:06:21.213609 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" event={"ID":"66a68296-c0ef-463c-8d6f-e6a970833180","Type":"ContainerStarted","Data":"ff56aecec95681d64e15a837bcf70d0b408ffb25ed36cc7e78952dcd15565200"} Oct 05 07:06:21 crc kubenswrapper[4935]: I1005 07:06:21.217632 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" event={"ID":"2a702a6b-c2f1-46ae-b6da-a99ff519a83f","Type":"ContainerStarted","Data":"c4bdd2f40d57c2cc26793b86315d5c62ee0c0df1729666774a87b1813da280de"} Oct 05 07:06:21 crc kubenswrapper[4935]: I1005 07:06:21.220957 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qxgvc" event={"ID":"7d9b8112-d308-439e-838d-410a3bc6bf18","Type":"ContainerStarted","Data":"76018590f4b8138f72c152c64f3a71e59a3036b0c62f27e6f99041f95f3e9fd0"} Oct 05 07:06:21 crc kubenswrapper[4935]: I1005 07:06:21.263659 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:21 crc kubenswrapper[4935]: I1005 07:06:21.319168 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:21 crc kubenswrapper[4935]: I1005 07:06:21.503458 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6qnhq"] Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.230459 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6qnhq" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="registry-server" containerID="cri-o://0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8" gracePeriod=2 Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.600046 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.732707 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m58z2\" (UniqueName: \"kubernetes.io/projected/089298bd-a478-4011-9320-1f78ed045035-kube-api-access-m58z2\") pod \"089298bd-a478-4011-9320-1f78ed045035\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.732748 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-utilities\") pod \"089298bd-a478-4011-9320-1f78ed045035\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.733840 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-utilities" (OuterVolumeSpecName: "utilities") pod "089298bd-a478-4011-9320-1f78ed045035" (UID: "089298bd-a478-4011-9320-1f78ed045035"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.733960 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-catalog-content\") pod \"089298bd-a478-4011-9320-1f78ed045035\" (UID: \"089298bd-a478-4011-9320-1f78ed045035\") " Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.745250 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089298bd-a478-4011-9320-1f78ed045035-kube-api-access-m58z2" (OuterVolumeSpecName: "kube-api-access-m58z2") pod "089298bd-a478-4011-9320-1f78ed045035" (UID: "089298bd-a478-4011-9320-1f78ed045035"). InnerVolumeSpecName "kube-api-access-m58z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.746936 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m58z2\" (UniqueName: \"kubernetes.io/projected/089298bd-a478-4011-9320-1f78ed045035-kube-api-access-m58z2\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.746963 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.823901 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "089298bd-a478-4011-9320-1f78ed045035" (UID: "089298bd-a478-4011-9320-1f78ed045035"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:23 crc kubenswrapper[4935]: I1005 07:06:23.848082 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089298bd-a478-4011-9320-1f78ed045035-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.239852 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" event={"ID":"406136d1-87c2-4e34-9941-38d84f0aaff4","Type":"ContainerStarted","Data":"7482311a107e70efefbefdb48313fc34999e8b4f8fb47544f92c89edae1ddc20"} Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.242939 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qxgvc" event={"ID":"7d9b8112-d308-439e-838d-410a3bc6bf18","Type":"ContainerStarted","Data":"e73564f0ad8da53e0f8979fc5c54ea68130e97f6bda9bb3c2dbf419423b4366f"} Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.243256 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.245066 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" event={"ID":"66a68296-c0ef-463c-8d6f-e6a970833180","Type":"ContainerStarted","Data":"82f179900447271457f743b626a98bc8e9e55153cdd0bf1de02cebf4feaa3196"} Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.245178 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.246916 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" event={"ID":"2a702a6b-c2f1-46ae-b6da-a99ff519a83f","Type":"ContainerStarted","Data":"2903e4e5c7936c8f81bab470c89d472196659b0402d2b5c4f78d1e893c8dc465"} Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.249317 4935 generic.go:334] "Generic (PLEG): container finished" podID="089298bd-a478-4011-9320-1f78ed045035" containerID="0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8" exitCode=0 Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.249359 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qnhq" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.249360 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qnhq" event={"ID":"089298bd-a478-4011-9320-1f78ed045035","Type":"ContainerDied","Data":"0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8"} Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.249410 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qnhq" event={"ID":"089298bd-a478-4011-9320-1f78ed045035","Type":"ContainerDied","Data":"f788774d33f3dd84d6e0a7f764d30924e94365527f97e7686c25f0619815737b"} Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.249428 4935 scope.go:117] "RemoveContainer" containerID="0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.264451 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-bl5mm" podStartSLOduration=3.090480308 podStartE2EDuration="6.264430961s" podCreationTimestamp="2025-10-05 07:06:18 +0000 UTC" firstStartedPulling="2025-10-05 07:06:20.001042521 +0000 UTC m=+813.883668981" lastFinishedPulling="2025-10-05 07:06:23.174993174 +0000 UTC m=+817.057619634" observedRunningTime="2025-10-05 07:06:24.254295111 +0000 UTC m=+818.136921571" watchObservedRunningTime="2025-10-05 07:06:24.264430961 +0000 UTC m=+818.147057421" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.275212 4935 scope.go:117] "RemoveContainer" containerID="d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.282223 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-qxgvc" podStartSLOduration=3.275563934 podStartE2EDuration="6.282209193s" podCreationTimestamp="2025-10-05 07:06:18 +0000 UTC" firstStartedPulling="2025-10-05 07:06:20.211498081 +0000 UTC m=+814.094124541" lastFinishedPulling="2025-10-05 07:06:23.21814333 +0000 UTC m=+817.100769800" observedRunningTime="2025-10-05 07:06:24.279196443 +0000 UTC m=+818.161822903" watchObservedRunningTime="2025-10-05 07:06:24.282209193 +0000 UTC m=+818.164835653" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.308475 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" podStartSLOduration=3.678188828 podStartE2EDuration="6.3084595s" podCreationTimestamp="2025-10-05 07:06:18 +0000 UTC" firstStartedPulling="2025-10-05 07:06:20.635547014 +0000 UTC m=+814.518173494" lastFinishedPulling="2025-10-05 07:06:23.265817706 +0000 UTC m=+817.148444166" observedRunningTime="2025-10-05 07:06:24.302299186 +0000 UTC m=+818.184925666" watchObservedRunningTime="2025-10-05 07:06:24.3084595 +0000 UTC m=+818.191085960" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.309464 4935 scope.go:117] "RemoveContainer" containerID="269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.323153 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6qnhq"] Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.326747 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6qnhq"] Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.328568 4935 scope.go:117] "RemoveContainer" containerID="0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8" Oct 05 07:06:24 crc kubenswrapper[4935]: E1005 07:06:24.328996 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8\": container with ID starting with 0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8 not found: ID does not exist" containerID="0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.329027 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8"} err="failed to get container status \"0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8\": rpc error: code = NotFound desc = could not find container \"0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8\": container with ID starting with 0110082e8a26422aa70d65c956bccbc2ebe716f763f3076a2bae834e45b5ded8 not found: ID does not exist" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.329048 4935 scope.go:117] "RemoveContainer" containerID="d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6" Oct 05 07:06:24 crc kubenswrapper[4935]: E1005 07:06:24.329214 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6\": container with ID starting with d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6 not found: ID does not exist" containerID="d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.329235 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6"} err="failed to get container status \"d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6\": rpc error: code = NotFound desc = could not find container \"d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6\": container with ID starting with d4c65f101be1b335901386ed12a355cb45328e74ae396cfe1290711814d884e6 not found: ID does not exist" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.329247 4935 scope.go:117] "RemoveContainer" containerID="269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335" Oct 05 07:06:24 crc kubenswrapper[4935]: E1005 07:06:24.329384 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335\": container with ID starting with 269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335 not found: ID does not exist" containerID="269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.329405 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335"} err="failed to get container status \"269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335\": rpc error: code = NotFound desc = could not find container \"269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335\": container with ID starting with 269a840b4b88ffab3dc31bcb6466b1236e8d5b349dfa165c67db61e9a7b38335 not found: ID does not exist" Oct 05 07:06:24 crc kubenswrapper[4935]: I1005 07:06:24.785276 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089298bd-a478-4011-9320-1f78ed045035" path="/var/lib/kubelet/pods/089298bd-a478-4011-9320-1f78ed045035/volumes" Oct 05 07:06:26 crc kubenswrapper[4935]: I1005 07:06:26.264310 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" event={"ID":"2a702a6b-c2f1-46ae-b6da-a99ff519a83f","Type":"ContainerStarted","Data":"0f2a89c58c53b3e176a9f84fecc0d23c973ff428562ce6b87ecb281741640539"} Oct 05 07:06:26 crc kubenswrapper[4935]: I1005 07:06:26.279851 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-blpj9" podStartSLOduration=3.414652607 podStartE2EDuration="8.279832321s" podCreationTimestamp="2025-10-05 07:06:18 +0000 UTC" firstStartedPulling="2025-10-05 07:06:20.532261681 +0000 UTC m=+814.414888131" lastFinishedPulling="2025-10-05 07:06:25.397441385 +0000 UTC m=+819.280067845" observedRunningTime="2025-10-05 07:06:26.277698554 +0000 UTC m=+820.160325014" watchObservedRunningTime="2025-10-05 07:06:26.279832321 +0000 UTC m=+820.162458781" Oct 05 07:06:29 crc kubenswrapper[4935]: I1005 07:06:29.410499 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:29 crc kubenswrapper[4935]: I1005 07:06:29.410842 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:29 crc kubenswrapper[4935]: I1005 07:06:29.415239 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:30 crc kubenswrapper[4935]: I1005 07:06:30.145328 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-qxgvc" Oct 05 07:06:30 crc kubenswrapper[4935]: I1005 07:06:30.291574 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-78c49f5cb6-zvfvk" Oct 05 07:06:30 crc kubenswrapper[4935]: I1005 07:06:30.378496 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-cbb4c"] Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.656189 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pg2ld"] Oct 05 07:06:37 crc kubenswrapper[4935]: E1005 07:06:37.658136 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="extract-content" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.658163 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="extract-content" Oct 05 07:06:37 crc kubenswrapper[4935]: E1005 07:06:37.658347 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="extract-utilities" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.658373 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="extract-utilities" Oct 05 07:06:37 crc kubenswrapper[4935]: E1005 07:06:37.658401 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="registry-server" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.658417 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="registry-server" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.666147 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="089298bd-a478-4011-9320-1f78ed045035" containerName="registry-server" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.667260 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.667356 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pg2ld"] Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.731101 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-catalog-content\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.731172 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-utilities\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.731227 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26nsz\" (UniqueName: \"kubernetes.io/projected/83caacc3-e6e5-4744-aa36-acd31de96e9e-kube-api-access-26nsz\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.832124 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-utilities\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.832466 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26nsz\" (UniqueName: \"kubernetes.io/projected/83caacc3-e6e5-4744-aa36-acd31de96e9e-kube-api-access-26nsz\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.832662 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-catalog-content\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.833150 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-utilities\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.833287 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-catalog-content\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.854167 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26nsz\" (UniqueName: \"kubernetes.io/projected/83caacc3-e6e5-4744-aa36-acd31de96e9e-kube-api-access-26nsz\") pod \"certified-operators-pg2ld\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:37 crc kubenswrapper[4935]: I1005 07:06:37.998596 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:38 crc kubenswrapper[4935]: I1005 07:06:38.456011 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pg2ld"] Oct 05 07:06:39 crc kubenswrapper[4935]: I1005 07:06:39.348133 4935 generic.go:334] "Generic (PLEG): container finished" podID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerID="6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de" exitCode=0 Oct 05 07:06:39 crc kubenswrapper[4935]: I1005 07:06:39.348285 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pg2ld" event={"ID":"83caacc3-e6e5-4744-aa36-acd31de96e9e","Type":"ContainerDied","Data":"6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de"} Oct 05 07:06:39 crc kubenswrapper[4935]: I1005 07:06:39.348489 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pg2ld" event={"ID":"83caacc3-e6e5-4744-aa36-acd31de96e9e","Type":"ContainerStarted","Data":"9f2521e105136c7a19670a9e2a31469c6d80557dc95ff20955f34d8e7d5d5b82"} Oct 05 07:06:40 crc kubenswrapper[4935]: I1005 07:06:40.240957 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-n8448" Oct 05 07:06:40 crc kubenswrapper[4935]: I1005 07:06:40.366477 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pg2ld" event={"ID":"83caacc3-e6e5-4744-aa36-acd31de96e9e","Type":"ContainerStarted","Data":"d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9"} Oct 05 07:06:41 crc kubenswrapper[4935]: I1005 07:06:41.373124 4935 generic.go:334] "Generic (PLEG): container finished" podID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerID="d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9" exitCode=0 Oct 05 07:06:41 crc kubenswrapper[4935]: I1005 07:06:41.373174 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pg2ld" event={"ID":"83caacc3-e6e5-4744-aa36-acd31de96e9e","Type":"ContainerDied","Data":"d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9"} Oct 05 07:06:42 crc kubenswrapper[4935]: I1005 07:06:42.380416 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pg2ld" event={"ID":"83caacc3-e6e5-4744-aa36-acd31de96e9e","Type":"ContainerStarted","Data":"e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0"} Oct 05 07:06:42 crc kubenswrapper[4935]: I1005 07:06:42.395759 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pg2ld" podStartSLOduration=2.909085769 podStartE2EDuration="5.395741096s" podCreationTimestamp="2025-10-05 07:06:37 +0000 UTC" firstStartedPulling="2025-10-05 07:06:39.350493672 +0000 UTC m=+833.233120162" lastFinishedPulling="2025-10-05 07:06:41.837149029 +0000 UTC m=+835.719775489" observedRunningTime="2025-10-05 07:06:42.395157521 +0000 UTC m=+836.277783991" watchObservedRunningTime="2025-10-05 07:06:42.395741096 +0000 UTC m=+836.278367556" Oct 05 07:06:47 crc kubenswrapper[4935]: I1005 07:06:47.999342 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:47 crc kubenswrapper[4935]: I1005 07:06:47.999910 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:48 crc kubenswrapper[4935]: I1005 07:06:48.064993 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:48 crc kubenswrapper[4935]: I1005 07:06:48.522678 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:48 crc kubenswrapper[4935]: I1005 07:06:48.564812 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pg2ld"] Oct 05 07:06:50 crc kubenswrapper[4935]: I1005 07:06:50.432715 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pg2ld" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerName="registry-server" containerID="cri-o://e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0" gracePeriod=2 Oct 05 07:06:50 crc kubenswrapper[4935]: I1005 07:06:50.899145 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:50 crc kubenswrapper[4935]: I1005 07:06:50.923000 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26nsz\" (UniqueName: \"kubernetes.io/projected/83caacc3-e6e5-4744-aa36-acd31de96e9e-kube-api-access-26nsz\") pod \"83caacc3-e6e5-4744-aa36-acd31de96e9e\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " Oct 05 07:06:50 crc kubenswrapper[4935]: I1005 07:06:50.923064 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-catalog-content\") pod \"83caacc3-e6e5-4744-aa36-acd31de96e9e\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " Oct 05 07:06:50 crc kubenswrapper[4935]: I1005 07:06:50.923135 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-utilities\") pod \"83caacc3-e6e5-4744-aa36-acd31de96e9e\" (UID: \"83caacc3-e6e5-4744-aa36-acd31de96e9e\") " Oct 05 07:06:50 crc kubenswrapper[4935]: I1005 07:06:50.924100 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-utilities" (OuterVolumeSpecName: "utilities") pod "83caacc3-e6e5-4744-aa36-acd31de96e9e" (UID: "83caacc3-e6e5-4744-aa36-acd31de96e9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:50 crc kubenswrapper[4935]: I1005 07:06:50.930534 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83caacc3-e6e5-4744-aa36-acd31de96e9e-kube-api-access-26nsz" (OuterVolumeSpecName: "kube-api-access-26nsz") pod "83caacc3-e6e5-4744-aa36-acd31de96e9e" (UID: "83caacc3-e6e5-4744-aa36-acd31de96e9e"). InnerVolumeSpecName "kube-api-access-26nsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:50 crc kubenswrapper[4935]: I1005 07:06:50.967136 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83caacc3-e6e5-4744-aa36-acd31de96e9e" (UID: "83caacc3-e6e5-4744-aa36-acd31de96e9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.024746 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26nsz\" (UniqueName: \"kubernetes.io/projected/83caacc3-e6e5-4744-aa36-acd31de96e9e-kube-api-access-26nsz\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.024791 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.024804 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83caacc3-e6e5-4744-aa36-acd31de96e9e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.439925 4935 generic.go:334] "Generic (PLEG): container finished" podID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerID="e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0" exitCode=0 Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.439961 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pg2ld" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.439993 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pg2ld" event={"ID":"83caacc3-e6e5-4744-aa36-acd31de96e9e","Type":"ContainerDied","Data":"e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0"} Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.440408 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pg2ld" event={"ID":"83caacc3-e6e5-4744-aa36-acd31de96e9e","Type":"ContainerDied","Data":"9f2521e105136c7a19670a9e2a31469c6d80557dc95ff20955f34d8e7d5d5b82"} Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.440432 4935 scope.go:117] "RemoveContainer" containerID="e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.459910 4935 scope.go:117] "RemoveContainer" containerID="d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.474439 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pg2ld"] Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.477851 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pg2ld"] Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.479543 4935 scope.go:117] "RemoveContainer" containerID="6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.513935 4935 scope.go:117] "RemoveContainer" containerID="e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0" Oct 05 07:06:51 crc kubenswrapper[4935]: E1005 07:06:51.514996 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0\": container with ID starting with e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0 not found: ID does not exist" containerID="e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.515042 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0"} err="failed to get container status \"e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0\": rpc error: code = NotFound desc = could not find container \"e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0\": container with ID starting with e58cab2bbe37fd999f3d02d65a3858446a64ddfc2a188a7cc269266abe7498e0 not found: ID does not exist" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.515067 4935 scope.go:117] "RemoveContainer" containerID="d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9" Oct 05 07:06:51 crc kubenswrapper[4935]: E1005 07:06:51.515555 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9\": container with ID starting with d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9 not found: ID does not exist" containerID="d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.515601 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9"} err="failed to get container status \"d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9\": rpc error: code = NotFound desc = could not find container \"d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9\": container with ID starting with d5408c80f312e81e13fc1c6df7ba9abf86f923c65e5b07fb3935f35b80de29f9 not found: ID does not exist" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.515620 4935 scope.go:117] "RemoveContainer" containerID="6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de" Oct 05 07:06:51 crc kubenswrapper[4935]: E1005 07:06:51.515881 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de\": container with ID starting with 6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de not found: ID does not exist" containerID="6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de" Oct 05 07:06:51 crc kubenswrapper[4935]: I1005 07:06:51.515937 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de"} err="failed to get container status \"6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de\": rpc error: code = NotFound desc = could not find container \"6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de\": container with ID starting with 6612b2ad5ed4a2a6b4bad420346285d02521b704e7a8b88f49c5416711fe89de not found: ID does not exist" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.761249 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf"] Oct 05 07:06:52 crc kubenswrapper[4935]: E1005 07:06:52.761604 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerName="extract-utilities" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.761626 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerName="extract-utilities" Oct 05 07:06:52 crc kubenswrapper[4935]: E1005 07:06:52.761655 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerName="registry-server" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.761669 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerName="registry-server" Oct 05 07:06:52 crc kubenswrapper[4935]: E1005 07:06:52.761693 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerName="extract-content" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.761707 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerName="extract-content" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.761950 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" containerName="registry-server" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.763413 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.766138 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.769385 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf"] Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.785332 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83caacc3-e6e5-4744-aa36-acd31de96e9e" path="/var/lib/kubelet/pods/83caacc3-e6e5-4744-aa36-acd31de96e9e/volumes" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.949214 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.949408 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhh6c\" (UniqueName: \"kubernetes.io/projected/a5685c69-6cce-44df-9252-84962c0cc76a-kube-api-access-vhh6c\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:52 crc kubenswrapper[4935]: I1005 07:06:52.949478 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:53 crc kubenswrapper[4935]: I1005 07:06:53.051227 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhh6c\" (UniqueName: \"kubernetes.io/projected/a5685c69-6cce-44df-9252-84962c0cc76a-kube-api-access-vhh6c\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:53 crc kubenswrapper[4935]: I1005 07:06:53.051305 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:53 crc kubenswrapper[4935]: I1005 07:06:53.051368 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:53 crc kubenswrapper[4935]: I1005 07:06:53.052370 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:53 crc kubenswrapper[4935]: I1005 07:06:53.052514 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:53 crc kubenswrapper[4935]: I1005 07:06:53.079265 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhh6c\" (UniqueName: \"kubernetes.io/projected/a5685c69-6cce-44df-9252-84962c0cc76a-kube-api-access-vhh6c\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:53 crc kubenswrapper[4935]: I1005 07:06:53.093505 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:53 crc kubenswrapper[4935]: I1005 07:06:53.515135 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf"] Oct 05 07:06:54 crc kubenswrapper[4935]: I1005 07:06:54.461862 4935 generic.go:334] "Generic (PLEG): container finished" podID="a5685c69-6cce-44df-9252-84962c0cc76a" containerID="b4e27c8f127d3c6ab3f90b9862ea90b4669e8a0013619dc0cefb2a2fc0cb5e09" exitCode=0 Oct 05 07:06:54 crc kubenswrapper[4935]: I1005 07:06:54.461962 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" event={"ID":"a5685c69-6cce-44df-9252-84962c0cc76a","Type":"ContainerDied","Data":"b4e27c8f127d3c6ab3f90b9862ea90b4669e8a0013619dc0cefb2a2fc0cb5e09"} Oct 05 07:06:54 crc kubenswrapper[4935]: I1005 07:06:54.462288 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" event={"ID":"a5685c69-6cce-44df-9252-84962c0cc76a","Type":"ContainerStarted","Data":"48fdc2eb1abd139749e54af82e717955e1a0b15fecf47e3cce829885729613bc"} Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.437091 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-cbb4c" podUID="ac0661f5-9317-459e-bfce-c2ff9058c319" containerName="console" containerID="cri-o://c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e" gracePeriod=15 Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.834005 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-cbb4c_ac0661f5-9317-459e-bfce-c2ff9058c319/console/0.log" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.834067 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.892446 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-console-config\") pod \"ac0661f5-9317-459e-bfce-c2ff9058c319\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.892803 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hgg5\" (UniqueName: \"kubernetes.io/projected/ac0661f5-9317-459e-bfce-c2ff9058c319-kube-api-access-4hgg5\") pod \"ac0661f5-9317-459e-bfce-c2ff9058c319\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.892842 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-service-ca\") pod \"ac0661f5-9317-459e-bfce-c2ff9058c319\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.892877 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-serving-cert\") pod \"ac0661f5-9317-459e-bfce-c2ff9058c319\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.892928 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-oauth-config\") pod \"ac0661f5-9317-459e-bfce-c2ff9058c319\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.892950 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-oauth-serving-cert\") pod \"ac0661f5-9317-459e-bfce-c2ff9058c319\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.893384 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-console-config" (OuterVolumeSpecName: "console-config") pod "ac0661f5-9317-459e-bfce-c2ff9058c319" (UID: "ac0661f5-9317-459e-bfce-c2ff9058c319"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.893711 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ac0661f5-9317-459e-bfce-c2ff9058c319" (UID: "ac0661f5-9317-459e-bfce-c2ff9058c319"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.893943 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-service-ca" (OuterVolumeSpecName: "service-ca") pod "ac0661f5-9317-459e-bfce-c2ff9058c319" (UID: "ac0661f5-9317-459e-bfce-c2ff9058c319"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.907429 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ac0661f5-9317-459e-bfce-c2ff9058c319" (UID: "ac0661f5-9317-459e-bfce-c2ff9058c319"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.907748 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ac0661f5-9317-459e-bfce-c2ff9058c319" (UID: "ac0661f5-9317-459e-bfce-c2ff9058c319"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.911342 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac0661f5-9317-459e-bfce-c2ff9058c319-kube-api-access-4hgg5" (OuterVolumeSpecName: "kube-api-access-4hgg5") pod "ac0661f5-9317-459e-bfce-c2ff9058c319" (UID: "ac0661f5-9317-459e-bfce-c2ff9058c319"). InnerVolumeSpecName "kube-api-access-4hgg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.993667 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-trusted-ca-bundle\") pod \"ac0661f5-9317-459e-bfce-c2ff9058c319\" (UID: \"ac0661f5-9317-459e-bfce-c2ff9058c319\") " Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.993919 4935 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.993937 4935 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.993948 4935 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-console-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.993959 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hgg5\" (UniqueName: \"kubernetes.io/projected/ac0661f5-9317-459e-bfce-c2ff9058c319-kube-api-access-4hgg5\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.993970 4935 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.993981 4935 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac0661f5-9317-459e-bfce-c2ff9058c319-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:55 crc kubenswrapper[4935]: I1005 07:06:55.994610 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ac0661f5-9317-459e-bfce-c2ff9058c319" (UID: "ac0661f5-9317-459e-bfce-c2ff9058c319"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.095010 4935 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac0661f5-9317-459e-bfce-c2ff9058c319-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.479194 4935 generic.go:334] "Generic (PLEG): container finished" podID="a5685c69-6cce-44df-9252-84962c0cc76a" containerID="0b993333c335f3310a8a984dc312987206bd6fadd414244682b86b383c5020f8" exitCode=0 Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.479547 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" event={"ID":"a5685c69-6cce-44df-9252-84962c0cc76a","Type":"ContainerDied","Data":"0b993333c335f3310a8a984dc312987206bd6fadd414244682b86b383c5020f8"} Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.481779 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-cbb4c_ac0661f5-9317-459e-bfce-c2ff9058c319/console/0.log" Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.481835 4935 generic.go:334] "Generic (PLEG): container finished" podID="ac0661f5-9317-459e-bfce-c2ff9058c319" containerID="c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e" exitCode=2 Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.481871 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cbb4c" event={"ID":"ac0661f5-9317-459e-bfce-c2ff9058c319","Type":"ContainerDied","Data":"c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e"} Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.481916 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cbb4c" event={"ID":"ac0661f5-9317-459e-bfce-c2ff9058c319","Type":"ContainerDied","Data":"0117757e0089912a92e1f34ca9432a4c2463541b9dd29cb07f60d5c862c29738"} Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.481943 4935 scope.go:117] "RemoveContainer" containerID="c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e" Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.482106 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cbb4c" Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.519645 4935 scope.go:117] "RemoveContainer" containerID="c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e" Oct 05 07:06:56 crc kubenswrapper[4935]: E1005 07:06:56.520497 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e\": container with ID starting with c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e not found: ID does not exist" containerID="c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e" Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.520586 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e"} err="failed to get container status \"c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e\": rpc error: code = NotFound desc = could not find container \"c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e\": container with ID starting with c482e1b8a70944cd6810a2f78942f52b82ad99fd7418e81370c5add31fa4956e not found: ID does not exist" Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.543733 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-cbb4c"] Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.550623 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-cbb4c"] Oct 05 07:06:56 crc kubenswrapper[4935]: I1005 07:06:56.786381 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac0661f5-9317-459e-bfce-c2ff9058c319" path="/var/lib/kubelet/pods/ac0661f5-9317-459e-bfce-c2ff9058c319/volumes" Oct 05 07:06:57 crc kubenswrapper[4935]: I1005 07:06:57.495410 4935 generic.go:334] "Generic (PLEG): container finished" podID="a5685c69-6cce-44df-9252-84962c0cc76a" containerID="527fc410a7b5669c807c639a07ecafbe3a3773ef4320f856b609bfa0b4c573a2" exitCode=0 Oct 05 07:06:57 crc kubenswrapper[4935]: I1005 07:06:57.495501 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" event={"ID":"a5685c69-6cce-44df-9252-84962c0cc76a","Type":"ContainerDied","Data":"527fc410a7b5669c807c639a07ecafbe3a3773ef4320f856b609bfa0b4c573a2"} Oct 05 07:06:58 crc kubenswrapper[4935]: I1005 07:06:58.840686 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.029043 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhh6c\" (UniqueName: \"kubernetes.io/projected/a5685c69-6cce-44df-9252-84962c0cc76a-kube-api-access-vhh6c\") pod \"a5685c69-6cce-44df-9252-84962c0cc76a\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.029382 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-bundle\") pod \"a5685c69-6cce-44df-9252-84962c0cc76a\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.029416 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-util\") pod \"a5685c69-6cce-44df-9252-84962c0cc76a\" (UID: \"a5685c69-6cce-44df-9252-84962c0cc76a\") " Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.030435 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-bundle" (OuterVolumeSpecName: "bundle") pod "a5685c69-6cce-44df-9252-84962c0cc76a" (UID: "a5685c69-6cce-44df-9252-84962c0cc76a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.034523 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5685c69-6cce-44df-9252-84962c0cc76a-kube-api-access-vhh6c" (OuterVolumeSpecName: "kube-api-access-vhh6c") pod "a5685c69-6cce-44df-9252-84962c0cc76a" (UID: "a5685c69-6cce-44df-9252-84962c0cc76a"). InnerVolumeSpecName "kube-api-access-vhh6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.061867 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-util" (OuterVolumeSpecName: "util") pod "a5685c69-6cce-44df-9252-84962c0cc76a" (UID: "a5685c69-6cce-44df-9252-84962c0cc76a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.130343 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhh6c\" (UniqueName: \"kubernetes.io/projected/a5685c69-6cce-44df-9252-84962c0cc76a-kube-api-access-vhh6c\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.130383 4935 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.130392 4935 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5685c69-6cce-44df-9252-84962c0cc76a-util\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.512663 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" event={"ID":"a5685c69-6cce-44df-9252-84962c0cc76a","Type":"ContainerDied","Data":"48fdc2eb1abd139749e54af82e717955e1a0b15fecf47e3cce829885729613bc"} Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.512722 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48fdc2eb1abd139749e54af82e717955e1a0b15fecf47e3cce829885729613bc" Oct 05 07:06:59 crc kubenswrapper[4935]: I1005 07:06:59.512784 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.478953 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl"] Oct 05 07:07:08 crc kubenswrapper[4935]: E1005 07:07:08.479618 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5685c69-6cce-44df-9252-84962c0cc76a" containerName="pull" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.479631 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5685c69-6cce-44df-9252-84962c0cc76a" containerName="pull" Oct 05 07:07:08 crc kubenswrapper[4935]: E1005 07:07:08.479647 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5685c69-6cce-44df-9252-84962c0cc76a" containerName="extract" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.479653 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5685c69-6cce-44df-9252-84962c0cc76a" containerName="extract" Oct 05 07:07:08 crc kubenswrapper[4935]: E1005 07:07:08.479665 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5685c69-6cce-44df-9252-84962c0cc76a" containerName="util" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.479670 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5685c69-6cce-44df-9252-84962c0cc76a" containerName="util" Oct 05 07:07:08 crc kubenswrapper[4935]: E1005 07:07:08.479679 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0661f5-9317-459e-bfce-c2ff9058c319" containerName="console" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.479684 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0661f5-9317-459e-bfce-c2ff9058c319" containerName="console" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.479774 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5685c69-6cce-44df-9252-84962c0cc76a" containerName="extract" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.479786 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac0661f5-9317-459e-bfce-c2ff9058c319" containerName="console" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.480167 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.488105 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.488110 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-dwrhd" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.488404 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.488120 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.489360 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.560601 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl"] Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.635810 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grzcv\" (UniqueName: \"kubernetes.io/projected/efba8319-0b0d-415d-bb5d-7d1381c85524-kube-api-access-grzcv\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.635902 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/efba8319-0b0d-415d-bb5d-7d1381c85524-apiservice-cert\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.635960 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/efba8319-0b0d-415d-bb5d-7d1381c85524-webhook-cert\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.736791 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/efba8319-0b0d-415d-bb5d-7d1381c85524-webhook-cert\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.736853 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grzcv\" (UniqueName: \"kubernetes.io/projected/efba8319-0b0d-415d-bb5d-7d1381c85524-kube-api-access-grzcv\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.736900 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/efba8319-0b0d-415d-bb5d-7d1381c85524-apiservice-cert\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.744491 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/efba8319-0b0d-415d-bb5d-7d1381c85524-apiservice-cert\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.746620 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/efba8319-0b0d-415d-bb5d-7d1381c85524-webhook-cert\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.757555 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grzcv\" (UniqueName: \"kubernetes.io/projected/efba8319-0b0d-415d-bb5d-7d1381c85524-kube-api-access-grzcv\") pod \"metallb-operator-controller-manager-944fbc97-lnpwl\" (UID: \"efba8319-0b0d-415d-bb5d-7d1381c85524\") " pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.796749 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.826061 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr"] Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.845093 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.850601 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.850866 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.856765 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-v74gd" Oct 05 07:07:08 crc kubenswrapper[4935]: I1005 07:07:08.857080 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr"] Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.053954 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mprq5\" (UniqueName: \"kubernetes.io/projected/7c48384d-5634-4320-b93b-b82d3eaf91ca-kube-api-access-mprq5\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.054397 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c48384d-5634-4320-b93b-b82d3eaf91ca-webhook-cert\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.054447 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c48384d-5634-4320-b93b-b82d3eaf91ca-apiservice-cert\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.155143 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mprq5\" (UniqueName: \"kubernetes.io/projected/7c48384d-5634-4320-b93b-b82d3eaf91ca-kube-api-access-mprq5\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.155218 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c48384d-5634-4320-b93b-b82d3eaf91ca-webhook-cert\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.155264 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c48384d-5634-4320-b93b-b82d3eaf91ca-apiservice-cert\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.158934 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c48384d-5634-4320-b93b-b82d3eaf91ca-apiservice-cert\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.159293 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c48384d-5634-4320-b93b-b82d3eaf91ca-webhook-cert\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.172801 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mprq5\" (UniqueName: \"kubernetes.io/projected/7c48384d-5634-4320-b93b-b82d3eaf91ca-kube-api-access-mprq5\") pod \"metallb-operator-webhook-server-659857b8df-5kfhr\" (UID: \"7c48384d-5634-4320-b93b-b82d3eaf91ca\") " pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.177323 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.278861 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl"] Oct 05 07:07:09 crc kubenswrapper[4935]: W1005 07:07:09.284865 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefba8319_0b0d_415d_bb5d_7d1381c85524.slice/crio-7dc6687bc344a04ff8a0c05c0432f97bffa5de9bf757d63ac1a72cbfe196a259 WatchSource:0}: Error finding container 7dc6687bc344a04ff8a0c05c0432f97bffa5de9bf757d63ac1a72cbfe196a259: Status 404 returned error can't find the container with id 7dc6687bc344a04ff8a0c05c0432f97bffa5de9bf757d63ac1a72cbfe196a259 Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.562621 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" event={"ID":"efba8319-0b0d-415d-bb5d-7d1381c85524","Type":"ContainerStarted","Data":"7dc6687bc344a04ff8a0c05c0432f97bffa5de9bf757d63ac1a72cbfe196a259"} Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.586330 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr"] Oct 05 07:07:09 crc kubenswrapper[4935]: W1005 07:07:09.590055 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c48384d_5634_4320_b93b_b82d3eaf91ca.slice/crio-4a291d177e578f25461f4dfaf359fce9ce94713db3c670eeb1397757614c4d26 WatchSource:0}: Error finding container 4a291d177e578f25461f4dfaf359fce9ce94713db3c670eeb1397757614c4d26: Status 404 returned error can't find the container with id 4a291d177e578f25461f4dfaf359fce9ce94713db3c670eeb1397757614c4d26 Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.900862 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q8v68"] Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.902443 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:09 crc kubenswrapper[4935]: I1005 07:07:09.910744 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8v68"] Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.064260 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-utilities\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.064338 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-catalog-content\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.064377 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtqpl\" (UniqueName: \"kubernetes.io/projected/c36a8347-c70c-494f-b63d-3672dbe60734-kube-api-access-xtqpl\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.165094 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-catalog-content\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.165177 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtqpl\" (UniqueName: \"kubernetes.io/projected/c36a8347-c70c-494f-b63d-3672dbe60734-kube-api-access-xtqpl\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.165219 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-utilities\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.165628 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-catalog-content\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.165739 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-utilities\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.197828 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtqpl\" (UniqueName: \"kubernetes.io/projected/c36a8347-c70c-494f-b63d-3672dbe60734-kube-api-access-xtqpl\") pod \"redhat-marketplace-q8v68\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.219548 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.489547 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8v68"] Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.574331 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8v68" event={"ID":"c36a8347-c70c-494f-b63d-3672dbe60734","Type":"ContainerStarted","Data":"d625291dd8f13e17b905bc855e5c8702c38f58fec957ce3a90a3e91d19479e76"} Oct 05 07:07:10 crc kubenswrapper[4935]: I1005 07:07:10.575549 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" event={"ID":"7c48384d-5634-4320-b93b-b82d3eaf91ca","Type":"ContainerStarted","Data":"4a291d177e578f25461f4dfaf359fce9ce94713db3c670eeb1397757614c4d26"} Oct 05 07:07:11 crc kubenswrapper[4935]: I1005 07:07:11.582976 4935 generic.go:334] "Generic (PLEG): container finished" podID="c36a8347-c70c-494f-b63d-3672dbe60734" containerID="03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f" exitCode=0 Oct 05 07:07:11 crc kubenswrapper[4935]: I1005 07:07:11.583194 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8v68" event={"ID":"c36a8347-c70c-494f-b63d-3672dbe60734","Type":"ContainerDied","Data":"03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f"} Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.289771 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.290197 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.602082 4935 generic.go:334] "Generic (PLEG): container finished" podID="c36a8347-c70c-494f-b63d-3672dbe60734" containerID="9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de" exitCode=0 Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.602154 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8v68" event={"ID":"c36a8347-c70c-494f-b63d-3672dbe60734","Type":"ContainerDied","Data":"9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de"} Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.603839 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" event={"ID":"7c48384d-5634-4320-b93b-b82d3eaf91ca","Type":"ContainerStarted","Data":"c8a9f830351224840102ad4c28a15f61fadc3bc5be2b299ed8ea5be6fd887728"} Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.603987 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.605023 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" event={"ID":"efba8319-0b0d-415d-bb5d-7d1381c85524","Type":"ContainerStarted","Data":"e42e2dd7a3dbb4966001c5e64b9e0206a2a4d3d6ed56750eb7152ceef529d0af"} Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.605170 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:14 crc kubenswrapper[4935]: I1005 07:07:14.641268 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" podStartSLOduration=1.6261581870000001 podStartE2EDuration="6.641249912s" podCreationTimestamp="2025-10-05 07:07:08 +0000 UTC" firstStartedPulling="2025-10-05 07:07:09.289602464 +0000 UTC m=+863.172228924" lastFinishedPulling="2025-10-05 07:07:14.304694189 +0000 UTC m=+868.187320649" observedRunningTime="2025-10-05 07:07:14.637957265 +0000 UTC m=+868.520583735" watchObservedRunningTime="2025-10-05 07:07:14.641249912 +0000 UTC m=+868.523876372" Oct 05 07:07:15 crc kubenswrapper[4935]: I1005 07:07:15.612726 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8v68" event={"ID":"c36a8347-c70c-494f-b63d-3672dbe60734","Type":"ContainerStarted","Data":"8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad"} Oct 05 07:07:15 crc kubenswrapper[4935]: I1005 07:07:15.638310 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q8v68" podStartSLOduration=3.137741796 podStartE2EDuration="6.638293427s" podCreationTimestamp="2025-10-05 07:07:09 +0000 UTC" firstStartedPulling="2025-10-05 07:07:11.584942606 +0000 UTC m=+865.467569066" lastFinishedPulling="2025-10-05 07:07:15.085494237 +0000 UTC m=+868.968120697" observedRunningTime="2025-10-05 07:07:15.634778505 +0000 UTC m=+869.517404965" watchObservedRunningTime="2025-10-05 07:07:15.638293427 +0000 UTC m=+869.520919887" Oct 05 07:07:15 crc kubenswrapper[4935]: I1005 07:07:15.638807 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" podStartSLOduration=2.906734106 podStartE2EDuration="7.638800511s" podCreationTimestamp="2025-10-05 07:07:08 +0000 UTC" firstStartedPulling="2025-10-05 07:07:09.593092054 +0000 UTC m=+863.475718514" lastFinishedPulling="2025-10-05 07:07:14.325158459 +0000 UTC m=+868.207784919" observedRunningTime="2025-10-05 07:07:14.659603236 +0000 UTC m=+868.542229696" watchObservedRunningTime="2025-10-05 07:07:15.638800511 +0000 UTC m=+869.521426971" Oct 05 07:07:20 crc kubenswrapper[4935]: I1005 07:07:20.220579 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:20 crc kubenswrapper[4935]: I1005 07:07:20.221068 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:20 crc kubenswrapper[4935]: I1005 07:07:20.273728 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:20 crc kubenswrapper[4935]: I1005 07:07:20.675830 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:22 crc kubenswrapper[4935]: I1005 07:07:22.502050 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8v68"] Oct 05 07:07:22 crc kubenswrapper[4935]: I1005 07:07:22.652045 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q8v68" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" containerName="registry-server" containerID="cri-o://8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad" gracePeriod=2 Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.013828 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.142534 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-catalog-content\") pod \"c36a8347-c70c-494f-b63d-3672dbe60734\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.142607 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtqpl\" (UniqueName: \"kubernetes.io/projected/c36a8347-c70c-494f-b63d-3672dbe60734-kube-api-access-xtqpl\") pod \"c36a8347-c70c-494f-b63d-3672dbe60734\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.142664 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-utilities\") pod \"c36a8347-c70c-494f-b63d-3672dbe60734\" (UID: \"c36a8347-c70c-494f-b63d-3672dbe60734\") " Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.143562 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-utilities" (OuterVolumeSpecName: "utilities") pod "c36a8347-c70c-494f-b63d-3672dbe60734" (UID: "c36a8347-c70c-494f-b63d-3672dbe60734"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.148962 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c36a8347-c70c-494f-b63d-3672dbe60734-kube-api-access-xtqpl" (OuterVolumeSpecName: "kube-api-access-xtqpl") pod "c36a8347-c70c-494f-b63d-3672dbe60734" (UID: "c36a8347-c70c-494f-b63d-3672dbe60734"). InnerVolumeSpecName "kube-api-access-xtqpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.158579 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c36a8347-c70c-494f-b63d-3672dbe60734" (UID: "c36a8347-c70c-494f-b63d-3672dbe60734"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.243823 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.243859 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c36a8347-c70c-494f-b63d-3672dbe60734-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.243874 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtqpl\" (UniqueName: \"kubernetes.io/projected/c36a8347-c70c-494f-b63d-3672dbe60734-kube-api-access-xtqpl\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.659294 4935 generic.go:334] "Generic (PLEG): container finished" podID="c36a8347-c70c-494f-b63d-3672dbe60734" containerID="8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad" exitCode=0 Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.659351 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8v68" event={"ID":"c36a8347-c70c-494f-b63d-3672dbe60734","Type":"ContainerDied","Data":"8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad"} Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.659367 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8v68" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.659392 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8v68" event={"ID":"c36a8347-c70c-494f-b63d-3672dbe60734","Type":"ContainerDied","Data":"d625291dd8f13e17b905bc855e5c8702c38f58fec957ce3a90a3e91d19479e76"} Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.659420 4935 scope.go:117] "RemoveContainer" containerID="8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.675835 4935 scope.go:117] "RemoveContainer" containerID="9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.698128 4935 scope.go:117] "RemoveContainer" containerID="03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.698880 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8v68"] Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.712662 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8v68"] Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.727838 4935 scope.go:117] "RemoveContainer" containerID="8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad" Oct 05 07:07:23 crc kubenswrapper[4935]: E1005 07:07:23.728259 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad\": container with ID starting with 8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad not found: ID does not exist" containerID="8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.728298 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad"} err="failed to get container status \"8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad\": rpc error: code = NotFound desc = could not find container \"8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad\": container with ID starting with 8005c95ef3d069f194705c452cb548b98aa4b327926a9ffca6db19e824ecb2ad not found: ID does not exist" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.728324 4935 scope.go:117] "RemoveContainer" containerID="9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de" Oct 05 07:07:23 crc kubenswrapper[4935]: E1005 07:07:23.728642 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de\": container with ID starting with 9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de not found: ID does not exist" containerID="9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.728694 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de"} err="failed to get container status \"9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de\": rpc error: code = NotFound desc = could not find container \"9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de\": container with ID starting with 9b8822bc5735ea564cf2c60fc6170ae82d4045690c179795a889c347090908de not found: ID does not exist" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.728728 4935 scope.go:117] "RemoveContainer" containerID="03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f" Oct 05 07:07:23 crc kubenswrapper[4935]: E1005 07:07:23.729084 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f\": container with ID starting with 03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f not found: ID does not exist" containerID="03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f" Oct 05 07:07:23 crc kubenswrapper[4935]: I1005 07:07:23.729114 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f"} err="failed to get container status \"03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f\": rpc error: code = NotFound desc = could not find container \"03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f\": container with ID starting with 03c659fe48f2a221b3f02d147bed0981633576126e776162694a4ca1eea43b9f not found: ID does not exist" Oct 05 07:07:24 crc kubenswrapper[4935]: I1005 07:07:24.785686 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" path="/var/lib/kubelet/pods/c36a8347-c70c-494f-b63d-3672dbe60734/volumes" Oct 05 07:07:29 crc kubenswrapper[4935]: I1005 07:07:29.198336 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-659857b8df-5kfhr" Oct 05 07:07:44 crc kubenswrapper[4935]: I1005 07:07:44.289248 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:07:44 crc kubenswrapper[4935]: I1005 07:07:44.289944 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:07:48 crc kubenswrapper[4935]: I1005 07:07:48.799164 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-944fbc97-lnpwl" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.470954 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw"] Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.471588 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" containerName="registry-server" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.471604 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" containerName="registry-server" Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.471623 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" containerName="extract-utilities" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.471632 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" containerName="extract-utilities" Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.471655 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" containerName="extract-content" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.471664 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" containerName="extract-content" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.471783 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c36a8347-c70c-494f-b63d-3672dbe60734" containerName="registry-server" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.472294 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.472463 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-xs9zt"] Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.478936 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.479135 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.479363 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-fkj44" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.482951 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw"] Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.486067 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.489293 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.564769 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-524b4"] Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.566243 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.569723 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.569852 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.570058 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-56pjf" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.570318 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.592090 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-4zp2r"] Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.593363 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.595348 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.606519 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-4zp2r"] Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.670865 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.670933 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrtnl\" (UniqueName: \"kubernetes.io/projected/4fb11792-cce4-4bbf-8df5-420e99305afa-kube-api-access-vrtnl\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.670959 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2875ff58-eeeb-477b-880f-efbfcf23ea8d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wfhfw\" (UID: \"2875ff58-eeeb-477b-880f-efbfcf23ea8d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.670997 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-cert\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671016 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rssrm\" (UniqueName: \"kubernetes.io/projected/2875ff58-eeeb-477b-880f-efbfcf23ea8d-kube-api-access-rssrm\") pod \"frr-k8s-webhook-server-64bf5d555-wfhfw\" (UID: \"2875ff58-eeeb-477b-880f-efbfcf23ea8d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671038 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-conf\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671057 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-reloader\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671085 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-metrics-certs\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671106 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/07951b31-51bf-45fd-b391-2195e76d97ce-metallb-excludel2\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671134 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqg4x\" (UniqueName: \"kubernetes.io/projected/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-kube-api-access-dqg4x\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671161 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-metrics-certs\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671194 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-sockets\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671212 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671238 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-startup\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671258 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics-certs\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.671280 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmklz\" (UniqueName: \"kubernetes.io/projected/07951b31-51bf-45fd-b391-2195e76d97ce-kube-api-access-kmklz\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772019 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-cert\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772078 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rssrm\" (UniqueName: \"kubernetes.io/projected/2875ff58-eeeb-477b-880f-efbfcf23ea8d-kube-api-access-rssrm\") pod \"frr-k8s-webhook-server-64bf5d555-wfhfw\" (UID: \"2875ff58-eeeb-477b-880f-efbfcf23ea8d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772101 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-conf\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772123 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-reloader\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772153 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-metrics-certs\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772178 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/07951b31-51bf-45fd-b391-2195e76d97ce-metallb-excludel2\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772207 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqg4x\" (UniqueName: \"kubernetes.io/projected/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-kube-api-access-dqg4x\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772234 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-metrics-certs\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772269 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-sockets\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772289 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772314 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-startup\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772332 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics-certs\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772355 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmklz\" (UniqueName: \"kubernetes.io/projected/07951b31-51bf-45fd-b391-2195e76d97ce-kube-api-access-kmklz\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772388 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrtnl\" (UniqueName: \"kubernetes.io/projected/4fb11792-cce4-4bbf-8df5-420e99305afa-kube-api-access-vrtnl\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772409 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.772436 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2875ff58-eeeb-477b-880f-efbfcf23ea8d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wfhfw\" (UID: \"2875ff58-eeeb-477b-880f-efbfcf23ea8d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.772568 4935 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.772625 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2875ff58-eeeb-477b-880f-efbfcf23ea8d-cert podName:2875ff58-eeeb-477b-880f-efbfcf23ea8d nodeName:}" failed. No retries permitted until 2025-10-05 07:07:50.272606203 +0000 UTC m=+904.155232663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2875ff58-eeeb-477b-880f-efbfcf23ea8d-cert") pod "frr-k8s-webhook-server-64bf5d555-wfhfw" (UID: "2875ff58-eeeb-477b-880f-efbfcf23ea8d") : secret "frr-k8s-webhook-server-cert" not found Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.772948 4935 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.772988 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-metrics-certs podName:07951b31-51bf-45fd-b391-2195e76d97ce nodeName:}" failed. No retries permitted until 2025-10-05 07:07:50.272977853 +0000 UTC m=+904.155604313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-metrics-certs") pod "speaker-524b4" (UID: "07951b31-51bf-45fd-b391-2195e76d97ce") : secret "speaker-certs-secret" not found Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.773188 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-conf\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.773333 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-sockets\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.773355 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-reloader\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.773427 4935 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.773472 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist podName:07951b31-51bf-45fd-b391-2195e76d97ce nodeName:}" failed. No retries permitted until 2025-10-05 07:07:50.273459306 +0000 UTC m=+904.156085766 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist") pod "speaker-524b4" (UID: "07951b31-51bf-45fd-b391-2195e76d97ce") : secret "metallb-memberlist" not found Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.773591 4935 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 05 07:07:49 crc kubenswrapper[4935]: E1005 07:07:49.773637 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics-certs podName:4fb11792-cce4-4bbf-8df5-420e99305afa nodeName:}" failed. No retries permitted until 2025-10-05 07:07:50.27362339 +0000 UTC m=+904.156249850 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics-certs") pod "frr-k8s-xs9zt" (UID: "4fb11792-cce4-4bbf-8df5-420e99305afa") : secret "frr-k8s-certs-secret" not found Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.773848 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/07951b31-51bf-45fd-b391-2195e76d97ce-metallb-excludel2\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.773845 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.774322 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4fb11792-cce4-4bbf-8df5-420e99305afa-frr-startup\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.779421 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-metrics-certs\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.779917 4935 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.786394 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-cert\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.792761 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rssrm\" (UniqueName: \"kubernetes.io/projected/2875ff58-eeeb-477b-880f-efbfcf23ea8d-kube-api-access-rssrm\") pod \"frr-k8s-webhook-server-64bf5d555-wfhfw\" (UID: \"2875ff58-eeeb-477b-880f-efbfcf23ea8d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.795661 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrtnl\" (UniqueName: \"kubernetes.io/projected/4fb11792-cce4-4bbf-8df5-420e99305afa-kube-api-access-vrtnl\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.800818 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqg4x\" (UniqueName: \"kubernetes.io/projected/a058bfb5-58b4-4ac1-885e-2e2748d2c4ee-kube-api-access-dqg4x\") pod \"controller-68d546b9d8-4zp2r\" (UID: \"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee\") " pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.808615 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmklz\" (UniqueName: \"kubernetes.io/projected/07951b31-51bf-45fd-b391-2195e76d97ce-kube-api-access-kmklz\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:49 crc kubenswrapper[4935]: I1005 07:07:49.912073 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.280700 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-metrics-certs\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.281286 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics-certs\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.281351 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.281392 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2875ff58-eeeb-477b-880f-efbfcf23ea8d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wfhfw\" (UID: \"2875ff58-eeeb-477b-880f-efbfcf23ea8d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:50 crc kubenswrapper[4935]: E1005 07:07:50.281548 4935 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 05 07:07:50 crc kubenswrapper[4935]: E1005 07:07:50.281635 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist podName:07951b31-51bf-45fd-b391-2195e76d97ce nodeName:}" failed. No retries permitted until 2025-10-05 07:07:51.281611808 +0000 UTC m=+905.164238268 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist") pod "speaker-524b4" (UID: "07951b31-51bf-45fd-b391-2195e76d97ce") : secret "metallb-memberlist" not found Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.284977 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4fb11792-cce4-4bbf-8df5-420e99305afa-metrics-certs\") pod \"frr-k8s-xs9zt\" (UID: \"4fb11792-cce4-4bbf-8df5-420e99305afa\") " pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.285138 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-metrics-certs\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.288150 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2875ff58-eeeb-477b-880f-efbfcf23ea8d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wfhfw\" (UID: \"2875ff58-eeeb-477b-880f-efbfcf23ea8d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.318373 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-4zp2r"] Oct 05 07:07:50 crc kubenswrapper[4935]: W1005 07:07:50.325720 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda058bfb5_58b4_4ac1_885e_2e2748d2c4ee.slice/crio-9ea75c4a35aec1ee0ac6e613776b74aacecdaf5ecf52f031d2016d90bf4ca20b WatchSource:0}: Error finding container 9ea75c4a35aec1ee0ac6e613776b74aacecdaf5ecf52f031d2016d90bf4ca20b: Status 404 returned error can't find the container with id 9ea75c4a35aec1ee0ac6e613776b74aacecdaf5ecf52f031d2016d90bf4ca20b Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.393710 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.399997 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.606314 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw"] Oct 05 07:07:50 crc kubenswrapper[4935]: W1005 07:07:50.614302 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2875ff58_eeeb_477b_880f_efbfcf23ea8d.slice/crio-389b20c636116f28428b909d7a4285526356ab6fef269916c6a9201bc978f1c4 WatchSource:0}: Error finding container 389b20c636116f28428b909d7a4285526356ab6fef269916c6a9201bc978f1c4: Status 404 returned error can't find the container with id 389b20c636116f28428b909d7a4285526356ab6fef269916c6a9201bc978f1c4 Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.816662 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerStarted","Data":"ea63031a19e6d3fddae8b84e9e8e5cc9153416cf99a86d43b18ffd5aae2050d7"} Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.818980 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" event={"ID":"2875ff58-eeeb-477b-880f-efbfcf23ea8d","Type":"ContainerStarted","Data":"389b20c636116f28428b909d7a4285526356ab6fef269916c6a9201bc978f1c4"} Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.820459 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4zp2r" event={"ID":"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee","Type":"ContainerStarted","Data":"dc60b7f8df6fa204cda38344a0d5c8079b7cdd3f48be5821bea959edcaca6bf3"} Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.820481 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4zp2r" event={"ID":"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee","Type":"ContainerStarted","Data":"1bfb00f972185a00e3b181197cae6eace2fb3b9c9cc4b5f635aa5f9fc35a4509"} Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.820492 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4zp2r" event={"ID":"a058bfb5-58b4-4ac1-885e-2e2748d2c4ee","Type":"ContainerStarted","Data":"9ea75c4a35aec1ee0ac6e613776b74aacecdaf5ecf52f031d2016d90bf4ca20b"} Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.820656 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:07:50 crc kubenswrapper[4935]: I1005 07:07:50.840179 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-4zp2r" podStartSLOduration=1.84016109 podStartE2EDuration="1.84016109s" podCreationTimestamp="2025-10-05 07:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:50.835767784 +0000 UTC m=+904.718394264" watchObservedRunningTime="2025-10-05 07:07:50.84016109 +0000 UTC m=+904.722787560" Oct 05 07:07:51 crc kubenswrapper[4935]: I1005 07:07:51.294105 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:51 crc kubenswrapper[4935]: I1005 07:07:51.299412 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/07951b31-51bf-45fd-b391-2195e76d97ce-memberlist\") pod \"speaker-524b4\" (UID: \"07951b31-51bf-45fd-b391-2195e76d97ce\") " pod="metallb-system/speaker-524b4" Oct 05 07:07:51 crc kubenswrapper[4935]: I1005 07:07:51.379466 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-524b4" Oct 05 07:07:51 crc kubenswrapper[4935]: W1005 07:07:51.402513 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07951b31_51bf_45fd_b391_2195e76d97ce.slice/crio-7039f7b0510f9e168b851efeadf8fdb33fd25af5df1fe490ccfe54d699bf9b5e WatchSource:0}: Error finding container 7039f7b0510f9e168b851efeadf8fdb33fd25af5df1fe490ccfe54d699bf9b5e: Status 404 returned error can't find the container with id 7039f7b0510f9e168b851efeadf8fdb33fd25af5df1fe490ccfe54d699bf9b5e Oct 05 07:07:51 crc kubenswrapper[4935]: I1005 07:07:51.832920 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-524b4" event={"ID":"07951b31-51bf-45fd-b391-2195e76d97ce","Type":"ContainerStarted","Data":"0ba183a2bb9cb4262a56b064e80b1a0b6ec1bd2775aaeeba6ba19723692596cb"} Oct 05 07:07:51 crc kubenswrapper[4935]: I1005 07:07:51.832963 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-524b4" event={"ID":"07951b31-51bf-45fd-b391-2195e76d97ce","Type":"ContainerStarted","Data":"7039f7b0510f9e168b851efeadf8fdb33fd25af5df1fe490ccfe54d699bf9b5e"} Oct 05 07:07:52 crc kubenswrapper[4935]: I1005 07:07:52.845748 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-524b4" event={"ID":"07951b31-51bf-45fd-b391-2195e76d97ce","Type":"ContainerStarted","Data":"bb61214cf31f492cfc554aea58a0b339c69ded1ab34e8110cea40dbc6363e0a8"} Oct 05 07:07:52 crc kubenswrapper[4935]: I1005 07:07:52.845910 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-524b4" Oct 05 07:07:52 crc kubenswrapper[4935]: I1005 07:07:52.861968 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-524b4" podStartSLOduration=3.86193138 podStartE2EDuration="3.86193138s" podCreationTimestamp="2025-10-05 07:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:52.859331072 +0000 UTC m=+906.741957532" watchObservedRunningTime="2025-10-05 07:07:52.86193138 +0000 UTC m=+906.744557840" Oct 05 07:07:57 crc kubenswrapper[4935]: I1005 07:07:57.879929 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" event={"ID":"2875ff58-eeeb-477b-880f-efbfcf23ea8d","Type":"ContainerStarted","Data":"05b81fec91ff056ec27aadefe72263714e54738845b1ca8cbcbb085da8ba9e03"} Oct 05 07:07:57 crc kubenswrapper[4935]: I1005 07:07:57.880637 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:07:57 crc kubenswrapper[4935]: I1005 07:07:57.882466 4935 generic.go:334] "Generic (PLEG): container finished" podID="4fb11792-cce4-4bbf-8df5-420e99305afa" containerID="1a48bd67348e04a9d6a444bf55d209e5310738e3650a1f545045e9cad12fac43" exitCode=0 Oct 05 07:07:57 crc kubenswrapper[4935]: I1005 07:07:57.882515 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerDied","Data":"1a48bd67348e04a9d6a444bf55d209e5310738e3650a1f545045e9cad12fac43"} Oct 05 07:07:57 crc kubenswrapper[4935]: I1005 07:07:57.899013 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" podStartSLOduration=2.156671413 podStartE2EDuration="8.898997937s" podCreationTimestamp="2025-10-05 07:07:49 +0000 UTC" firstStartedPulling="2025-10-05 07:07:50.616721722 +0000 UTC m=+904.499348182" lastFinishedPulling="2025-10-05 07:07:57.359048256 +0000 UTC m=+911.241674706" observedRunningTime="2025-10-05 07:07:57.894991811 +0000 UTC m=+911.777618261" watchObservedRunningTime="2025-10-05 07:07:57.898997937 +0000 UTC m=+911.781624397" Oct 05 07:07:58 crc kubenswrapper[4935]: I1005 07:07:58.889991 4935 generic.go:334] "Generic (PLEG): container finished" podID="4fb11792-cce4-4bbf-8df5-420e99305afa" containerID="fedc8c3bf88f775cc80c8868ec3048a1ab88784e58ead8e024c20e4e135e0a53" exitCode=0 Oct 05 07:07:58 crc kubenswrapper[4935]: I1005 07:07:58.890088 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerDied","Data":"fedc8c3bf88f775cc80c8868ec3048a1ab88784e58ead8e024c20e4e135e0a53"} Oct 05 07:07:59 crc kubenswrapper[4935]: I1005 07:07:59.897380 4935 generic.go:334] "Generic (PLEG): container finished" podID="4fb11792-cce4-4bbf-8df5-420e99305afa" containerID="828a3e592d744db9783f281aee674ff273a65681ad9fc1e698ae0cdd161f6db6" exitCode=0 Oct 05 07:07:59 crc kubenswrapper[4935]: I1005 07:07:59.897477 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerDied","Data":"828a3e592d744db9783f281aee674ff273a65681ad9fc1e698ae0cdd161f6db6"} Oct 05 07:08:00 crc kubenswrapper[4935]: I1005 07:08:00.906418 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerStarted","Data":"52c48581e9665a66c83c387c17c046ad253c975da5a0a3cbecc84b0b4d6f0266"} Oct 05 07:08:00 crc kubenswrapper[4935]: I1005 07:08:00.906457 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerStarted","Data":"dd577a9d03b683260c16677c2114e9c639bf26098da9b571ea721db7ce04e671"} Oct 05 07:08:00 crc kubenswrapper[4935]: I1005 07:08:00.906467 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerStarted","Data":"b0c070176b16381680753b86e26d13c228d2308c3c6cef2c7c5216deabf379b1"} Oct 05 07:08:00 crc kubenswrapper[4935]: I1005 07:08:00.906475 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerStarted","Data":"8132d2fee2ea787e9c8f8459493e2b616d241ad8e2b643629558f6a99299a81d"} Oct 05 07:08:01 crc kubenswrapper[4935]: I1005 07:08:01.383160 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-524b4" Oct 05 07:08:01 crc kubenswrapper[4935]: I1005 07:08:01.916139 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerStarted","Data":"112e19e458b856af92c22e638ea7989d4df72cd0260d565f829a8fc51cb98302"} Oct 05 07:08:01 crc kubenswrapper[4935]: I1005 07:08:01.916183 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xs9zt" event={"ID":"4fb11792-cce4-4bbf-8df5-420e99305afa","Type":"ContainerStarted","Data":"f090fed78b5fc42efd98962bb31d7b44fff4b96b2bbcbc55872ef25383747b32"} Oct 05 07:08:01 crc kubenswrapper[4935]: I1005 07:08:01.916315 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:08:01 crc kubenswrapper[4935]: I1005 07:08:01.938476 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-xs9zt" podStartSLOduration=6.117861274 podStartE2EDuration="12.938460063s" podCreationTimestamp="2025-10-05 07:07:49 +0000 UTC" firstStartedPulling="2025-10-05 07:07:50.519945788 +0000 UTC m=+904.402572248" lastFinishedPulling="2025-10-05 07:07:57.340544577 +0000 UTC m=+911.223171037" observedRunningTime="2025-10-05 07:08:01.934757755 +0000 UTC m=+915.817384235" watchObservedRunningTime="2025-10-05 07:08:01.938460063 +0000 UTC m=+915.821086523" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.637500 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k"] Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.638666 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.640630 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.656119 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.656190 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxm78\" (UniqueName: \"kubernetes.io/projected/5e538f9b-2152-463f-b3f9-f4a33f000688-kube-api-access-vxm78\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.656436 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.729068 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k"] Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.757795 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxm78\" (UniqueName: \"kubernetes.io/projected/5e538f9b-2152-463f-b3f9-f4a33f000688-kube-api-access-vxm78\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.757903 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.758019 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.758330 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.758553 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.780056 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxm78\" (UniqueName: \"kubernetes.io/projected/5e538f9b-2152-463f-b3f9-f4a33f000688-kube-api-access-vxm78\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:02 crc kubenswrapper[4935]: I1005 07:08:02.952270 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:03 crc kubenswrapper[4935]: I1005 07:08:03.360335 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k"] Oct 05 07:08:03 crc kubenswrapper[4935]: W1005 07:08:03.366749 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e538f9b_2152_463f_b3f9_f4a33f000688.slice/crio-1b4c764d5a9082201d25705bde991a949d4d8613dbb4c15ccdf341d28ee7827f WatchSource:0}: Error finding container 1b4c764d5a9082201d25705bde991a949d4d8613dbb4c15ccdf341d28ee7827f: Status 404 returned error can't find the container with id 1b4c764d5a9082201d25705bde991a949d4d8613dbb4c15ccdf341d28ee7827f Oct 05 07:08:03 crc kubenswrapper[4935]: I1005 07:08:03.926377 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" event={"ID":"5e538f9b-2152-463f-b3f9-f4a33f000688","Type":"ContainerStarted","Data":"1b4c764d5a9082201d25705bde991a949d4d8613dbb4c15ccdf341d28ee7827f"} Oct 05 07:08:04 crc kubenswrapper[4935]: I1005 07:08:04.935073 4935 generic.go:334] "Generic (PLEG): container finished" podID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerID="08d03eab80e6e717f3e8eb5a6a3c5676f5aee897afc066f779f3a4eaaa4d46fb" exitCode=0 Oct 05 07:08:04 crc kubenswrapper[4935]: I1005 07:08:04.935135 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" event={"ID":"5e538f9b-2152-463f-b3f9-f4a33f000688","Type":"ContainerDied","Data":"08d03eab80e6e717f3e8eb5a6a3c5676f5aee897afc066f779f3a4eaaa4d46fb"} Oct 05 07:08:05 crc kubenswrapper[4935]: I1005 07:08:05.400378 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:08:05 crc kubenswrapper[4935]: I1005 07:08:05.437879 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:08:08 crc kubenswrapper[4935]: I1005 07:08:08.958422 4935 generic.go:334] "Generic (PLEG): container finished" podID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerID="ae90eee3ba6976386ea5370c9130c164ea210e24688c8921b08f219e929b5405" exitCode=0 Oct 05 07:08:08 crc kubenswrapper[4935]: I1005 07:08:08.958524 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" event={"ID":"5e538f9b-2152-463f-b3f9-f4a33f000688","Type":"ContainerDied","Data":"ae90eee3ba6976386ea5370c9130c164ea210e24688c8921b08f219e929b5405"} Oct 05 07:08:09 crc kubenswrapper[4935]: I1005 07:08:09.917134 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-4zp2r" Oct 05 07:08:09 crc kubenswrapper[4935]: I1005 07:08:09.967103 4935 generic.go:334] "Generic (PLEG): container finished" podID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerID="4dd34040e68604b20691efc6ef24cabfbd721ed21344a582d17a1604cf649a2e" exitCode=0 Oct 05 07:08:09 crc kubenswrapper[4935]: I1005 07:08:09.967162 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" event={"ID":"5e538f9b-2152-463f-b3f9-f4a33f000688","Type":"ContainerDied","Data":"4dd34040e68604b20691efc6ef24cabfbd721ed21344a582d17a1604cf649a2e"} Oct 05 07:08:10 crc kubenswrapper[4935]: I1005 07:08:10.397860 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wfhfw" Oct 05 07:08:10 crc kubenswrapper[4935]: I1005 07:08:10.402326 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-xs9zt" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.231036 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.281651 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-util\") pod \"5e538f9b-2152-463f-b3f9-f4a33f000688\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.281883 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxm78\" (UniqueName: \"kubernetes.io/projected/5e538f9b-2152-463f-b3f9-f4a33f000688-kube-api-access-vxm78\") pod \"5e538f9b-2152-463f-b3f9-f4a33f000688\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.282172 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-bundle\") pod \"5e538f9b-2152-463f-b3f9-f4a33f000688\" (UID: \"5e538f9b-2152-463f-b3f9-f4a33f000688\") " Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.283221 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-bundle" (OuterVolumeSpecName: "bundle") pod "5e538f9b-2152-463f-b3f9-f4a33f000688" (UID: "5e538f9b-2152-463f-b3f9-f4a33f000688"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.288439 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e538f9b-2152-463f-b3f9-f4a33f000688-kube-api-access-vxm78" (OuterVolumeSpecName: "kube-api-access-vxm78") pod "5e538f9b-2152-463f-b3f9-f4a33f000688" (UID: "5e538f9b-2152-463f-b3f9-f4a33f000688"). InnerVolumeSpecName "kube-api-access-vxm78". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.319270 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-util" (OuterVolumeSpecName: "util") pod "5e538f9b-2152-463f-b3f9-f4a33f000688" (UID: "5e538f9b-2152-463f-b3f9-f4a33f000688"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.383996 4935 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.384478 4935 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e538f9b-2152-463f-b3f9-f4a33f000688-util\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.384493 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxm78\" (UniqueName: \"kubernetes.io/projected/5e538f9b-2152-463f-b3f9-f4a33f000688-kube-api-access-vxm78\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.981374 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" event={"ID":"5e538f9b-2152-463f-b3f9-f4a33f000688","Type":"ContainerDied","Data":"1b4c764d5a9082201d25705bde991a949d4d8613dbb4c15ccdf341d28ee7827f"} Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.981418 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b4c764d5a9082201d25705bde991a949d4d8613dbb4c15ccdf341d28ee7827f" Oct 05 07:08:11 crc kubenswrapper[4935]: I1005 07:08:11.981566 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k" Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.289212 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.289303 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.289352 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.289986 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"52b267e7f2c1f92355886ddbf4cef30a218042be6e418507f266c72f41da3c70"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.290046 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://52b267e7f2c1f92355886ddbf4cef30a218042be6e418507f266c72f41da3c70" gracePeriod=600 Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.998822 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="52b267e7f2c1f92355886ddbf4cef30a218042be6e418507f266c72f41da3c70" exitCode=0 Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.998936 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"52b267e7f2c1f92355886ddbf4cef30a218042be6e418507f266c72f41da3c70"} Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.999486 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"be9c0311e869ecdb94c82f2824221de74d838dab99b4132c5a117c8e772b2a41"} Oct 05 07:08:14 crc kubenswrapper[4935]: I1005 07:08:14.999513 4935 scope.go:117] "RemoveContainer" containerID="16ac36f74f68eb8c2c9e34c4f313b94ebe39f5ac0b6b53af3d6a75adcbd15fbc" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.940434 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877"] Oct 05 07:08:15 crc kubenswrapper[4935]: E1005 07:08:15.941138 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerName="extract" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.941155 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerName="extract" Oct 05 07:08:15 crc kubenswrapper[4935]: E1005 07:08:15.941178 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerName="pull" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.941184 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerName="pull" Oct 05 07:08:15 crc kubenswrapper[4935]: E1005 07:08:15.941199 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerName="util" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.941208 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerName="util" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.941375 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e538f9b-2152-463f-b3f9-f4a33f000688" containerName="extract" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.941998 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.944337 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.944706 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.944713 4935 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-m59dj" Oct 05 07:08:15 crc kubenswrapper[4935]: I1005 07:08:15.971735 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877"] Oct 05 07:08:16 crc kubenswrapper[4935]: I1005 07:08:16.046411 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slmw6\" (UniqueName: \"kubernetes.io/projected/a3c90fbf-845c-4728-957f-024943ea52f7-kube-api-access-slmw6\") pod \"cert-manager-operator-controller-manager-57cd46d6d-h8877\" (UID: \"a3c90fbf-845c-4728-957f-024943ea52f7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877" Oct 05 07:08:16 crc kubenswrapper[4935]: I1005 07:08:16.147987 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slmw6\" (UniqueName: \"kubernetes.io/projected/a3c90fbf-845c-4728-957f-024943ea52f7-kube-api-access-slmw6\") pod \"cert-manager-operator-controller-manager-57cd46d6d-h8877\" (UID: \"a3c90fbf-845c-4728-957f-024943ea52f7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877" Oct 05 07:08:16 crc kubenswrapper[4935]: I1005 07:08:16.172740 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slmw6\" (UniqueName: \"kubernetes.io/projected/a3c90fbf-845c-4728-957f-024943ea52f7-kube-api-access-slmw6\") pod \"cert-manager-operator-controller-manager-57cd46d6d-h8877\" (UID: \"a3c90fbf-845c-4728-957f-024943ea52f7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877" Oct 05 07:08:16 crc kubenswrapper[4935]: I1005 07:08:16.258203 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877" Oct 05 07:08:16 crc kubenswrapper[4935]: I1005 07:08:16.653701 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877"] Oct 05 07:08:17 crc kubenswrapper[4935]: I1005 07:08:17.028027 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877" event={"ID":"a3c90fbf-845c-4728-957f-024943ea52f7","Type":"ContainerStarted","Data":"9394bba2e473323cdddd92fce51a3dcaf72fe7bf42429fef0a2c102910769382"} Oct 05 07:08:24 crc kubenswrapper[4935]: I1005 07:08:24.072592 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877" event={"ID":"a3c90fbf-845c-4728-957f-024943ea52f7","Type":"ContainerStarted","Data":"c81da09e722ceee94594db3e6421f2c8154caedf6a6b591be6f2083b0f5a0b1d"} Oct 05 07:08:24 crc kubenswrapper[4935]: I1005 07:08:24.090834 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-h8877" podStartSLOduration=2.125180635 podStartE2EDuration="9.090814274s" podCreationTimestamp="2025-10-05 07:08:15 +0000 UTC" firstStartedPulling="2025-10-05 07:08:16.655878263 +0000 UTC m=+930.538504733" lastFinishedPulling="2025-10-05 07:08:23.621511902 +0000 UTC m=+937.504138372" observedRunningTime="2025-10-05 07:08:24.089266973 +0000 UTC m=+937.971893433" watchObservedRunningTime="2025-10-05 07:08:24.090814274 +0000 UTC m=+937.973440744" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.611565 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-rj6tr"] Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.612648 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.616222 4935 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-q6bqv" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.616553 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.616742 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.620565 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-rj6tr"] Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.719750 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0622bcbf-485e-48f2-8d5f-a359620318c4-bound-sa-token\") pod \"cert-manager-webhook-d969966f-rj6tr\" (UID: \"0622bcbf-485e-48f2-8d5f-a359620318c4\") " pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.719827 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h2rj\" (UniqueName: \"kubernetes.io/projected/0622bcbf-485e-48f2-8d5f-a359620318c4-kube-api-access-2h2rj\") pod \"cert-manager-webhook-d969966f-rj6tr\" (UID: \"0622bcbf-485e-48f2-8d5f-a359620318c4\") " pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.820570 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h2rj\" (UniqueName: \"kubernetes.io/projected/0622bcbf-485e-48f2-8d5f-a359620318c4-kube-api-access-2h2rj\") pod \"cert-manager-webhook-d969966f-rj6tr\" (UID: \"0622bcbf-485e-48f2-8d5f-a359620318c4\") " pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.820673 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0622bcbf-485e-48f2-8d5f-a359620318c4-bound-sa-token\") pod \"cert-manager-webhook-d969966f-rj6tr\" (UID: \"0622bcbf-485e-48f2-8d5f-a359620318c4\") " pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.851835 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h2rj\" (UniqueName: \"kubernetes.io/projected/0622bcbf-485e-48f2-8d5f-a359620318c4-kube-api-access-2h2rj\") pod \"cert-manager-webhook-d969966f-rj6tr\" (UID: \"0622bcbf-485e-48f2-8d5f-a359620318c4\") " pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.857562 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0622bcbf-485e-48f2-8d5f-a359620318c4-bound-sa-token\") pod \"cert-manager-webhook-d969966f-rj6tr\" (UID: \"0622bcbf-485e-48f2-8d5f-a359620318c4\") " pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:28 crc kubenswrapper[4935]: I1005 07:08:28.931873 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:29 crc kubenswrapper[4935]: I1005 07:08:29.346666 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-rj6tr"] Oct 05 07:08:29 crc kubenswrapper[4935]: W1005 07:08:29.351394 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0622bcbf_485e_48f2_8d5f_a359620318c4.slice/crio-797eada08bddc6250de081ca17b6248420c6f187d0a95fea3d49dd1300d64de3 WatchSource:0}: Error finding container 797eada08bddc6250de081ca17b6248420c6f187d0a95fea3d49dd1300d64de3: Status 404 returned error can't find the container with id 797eada08bddc6250de081ca17b6248420c6f187d0a95fea3d49dd1300d64de3 Oct 05 07:08:30 crc kubenswrapper[4935]: I1005 07:08:30.108128 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" event={"ID":"0622bcbf-485e-48f2-8d5f-a359620318c4","Type":"ContainerStarted","Data":"797eada08bddc6250de081ca17b6248420c6f187d0a95fea3d49dd1300d64de3"} Oct 05 07:08:30 crc kubenswrapper[4935]: I1005 07:08:30.951586 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk"] Oct 05 07:08:30 crc kubenswrapper[4935]: I1005 07:08:30.952475 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" Oct 05 07:08:30 crc kubenswrapper[4935]: I1005 07:08:30.960692 4935 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-s6sfg" Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.028614 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk"] Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.048383 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4qhh\" (UniqueName: \"kubernetes.io/projected/42349445-26e8-4b37-b947-9fe64536e3ab-kube-api-access-j4qhh\") pod \"cert-manager-cainjector-7d9f95dbf-c6tlk\" (UID: \"42349445-26e8-4b37-b947-9fe64536e3ab\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.048433 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42349445-26e8-4b37-b947-9fe64536e3ab-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-c6tlk\" (UID: \"42349445-26e8-4b37-b947-9fe64536e3ab\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.149676 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4qhh\" (UniqueName: \"kubernetes.io/projected/42349445-26e8-4b37-b947-9fe64536e3ab-kube-api-access-j4qhh\") pod \"cert-manager-cainjector-7d9f95dbf-c6tlk\" (UID: \"42349445-26e8-4b37-b947-9fe64536e3ab\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.149724 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42349445-26e8-4b37-b947-9fe64536e3ab-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-c6tlk\" (UID: \"42349445-26e8-4b37-b947-9fe64536e3ab\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.168491 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42349445-26e8-4b37-b947-9fe64536e3ab-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-c6tlk\" (UID: \"42349445-26e8-4b37-b947-9fe64536e3ab\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.172605 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4qhh\" (UniqueName: \"kubernetes.io/projected/42349445-26e8-4b37-b947-9fe64536e3ab-kube-api-access-j4qhh\") pod \"cert-manager-cainjector-7d9f95dbf-c6tlk\" (UID: \"42349445-26e8-4b37-b947-9fe64536e3ab\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.270173 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" Oct 05 07:08:31 crc kubenswrapper[4935]: I1005 07:08:31.674371 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk"] Oct 05 07:08:31 crc kubenswrapper[4935]: W1005 07:08:31.685752 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42349445_26e8_4b37_b947_9fe64536e3ab.slice/crio-08ee81f5ffe8e6e72903c01c8ded1e1c22808be65228030eae8c0f05d296eded WatchSource:0}: Error finding container 08ee81f5ffe8e6e72903c01c8ded1e1c22808be65228030eae8c0f05d296eded: Status 404 returned error can't find the container with id 08ee81f5ffe8e6e72903c01c8ded1e1c22808be65228030eae8c0f05d296eded Oct 05 07:08:32 crc kubenswrapper[4935]: I1005 07:08:32.118380 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" event={"ID":"42349445-26e8-4b37-b947-9fe64536e3ab","Type":"ContainerStarted","Data":"08ee81f5ffe8e6e72903c01c8ded1e1c22808be65228030eae8c0f05d296eded"} Oct 05 07:08:34 crc kubenswrapper[4935]: I1005 07:08:34.131220 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" event={"ID":"0622bcbf-485e-48f2-8d5f-a359620318c4","Type":"ContainerStarted","Data":"a6fc9ab92eabd4863e85d76cc2039b5766fb53d799063adcf42f437397e8b71e"} Oct 05 07:08:34 crc kubenswrapper[4935]: I1005 07:08:34.131593 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:34 crc kubenswrapper[4935]: I1005 07:08:34.132643 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" event={"ID":"42349445-26e8-4b37-b947-9fe64536e3ab","Type":"ContainerStarted","Data":"3f407f16968a9496dd6eadf3bd0e73c9c7ab68e5695cb3439ad5343c4d13f3d0"} Oct 05 07:08:34 crc kubenswrapper[4935]: I1005 07:08:34.152275 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" podStartSLOduration=1.966110414 podStartE2EDuration="6.152255781s" podCreationTimestamp="2025-10-05 07:08:28 +0000 UTC" firstStartedPulling="2025-10-05 07:08:29.354110438 +0000 UTC m=+943.236736898" lastFinishedPulling="2025-10-05 07:08:33.540255775 +0000 UTC m=+947.422882265" observedRunningTime="2025-10-05 07:08:34.14767407 +0000 UTC m=+948.030300530" watchObservedRunningTime="2025-10-05 07:08:34.152255781 +0000 UTC m=+948.034882241" Oct 05 07:08:34 crc kubenswrapper[4935]: I1005 07:08:34.166802 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-c6tlk" podStartSLOduration=2.332650225 podStartE2EDuration="4.166776565s" podCreationTimestamp="2025-10-05 07:08:30 +0000 UTC" firstStartedPulling="2025-10-05 07:08:31.688901309 +0000 UTC m=+945.571527769" lastFinishedPulling="2025-10-05 07:08:33.523027609 +0000 UTC m=+947.405654109" observedRunningTime="2025-10-05 07:08:34.164187017 +0000 UTC m=+948.046813497" watchObservedRunningTime="2025-10-05 07:08:34.166776565 +0000 UTC m=+948.049403055" Oct 05 07:08:38 crc kubenswrapper[4935]: I1005 07:08:38.935512 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-rj6tr" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.076419 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-82msx"] Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.078030 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-82msx" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.082694 4935 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-49p7f" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.088488 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-82msx"] Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.184123 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/171c85a1-08f1-481c-85fd-e60317147ad2-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-82msx\" (UID: \"171c85a1-08f1-481c-85fd-e60317147ad2\") " pod="cert-manager/cert-manager-7d4cc89fcb-82msx" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.184311 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccjgq\" (UniqueName: \"kubernetes.io/projected/171c85a1-08f1-481c-85fd-e60317147ad2-kube-api-access-ccjgq\") pod \"cert-manager-7d4cc89fcb-82msx\" (UID: \"171c85a1-08f1-481c-85fd-e60317147ad2\") " pod="cert-manager/cert-manager-7d4cc89fcb-82msx" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.287820 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccjgq\" (UniqueName: \"kubernetes.io/projected/171c85a1-08f1-481c-85fd-e60317147ad2-kube-api-access-ccjgq\") pod \"cert-manager-7d4cc89fcb-82msx\" (UID: \"171c85a1-08f1-481c-85fd-e60317147ad2\") " pod="cert-manager/cert-manager-7d4cc89fcb-82msx" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.287885 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/171c85a1-08f1-481c-85fd-e60317147ad2-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-82msx\" (UID: \"171c85a1-08f1-481c-85fd-e60317147ad2\") " pod="cert-manager/cert-manager-7d4cc89fcb-82msx" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.304991 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/171c85a1-08f1-481c-85fd-e60317147ad2-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-82msx\" (UID: \"171c85a1-08f1-481c-85fd-e60317147ad2\") " pod="cert-manager/cert-manager-7d4cc89fcb-82msx" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.309178 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccjgq\" (UniqueName: \"kubernetes.io/projected/171c85a1-08f1-481c-85fd-e60317147ad2-kube-api-access-ccjgq\") pod \"cert-manager-7d4cc89fcb-82msx\" (UID: \"171c85a1-08f1-481c-85fd-e60317147ad2\") " pod="cert-manager/cert-manager-7d4cc89fcb-82msx" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.420601 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-82msx" Oct 05 07:08:47 crc kubenswrapper[4935]: I1005 07:08:47.790600 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-82msx"] Oct 05 07:08:48 crc kubenswrapper[4935]: I1005 07:08:48.217188 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-82msx" event={"ID":"171c85a1-08f1-481c-85fd-e60317147ad2","Type":"ContainerStarted","Data":"1e41130243765fc4ac9e4e2d3650bb6a6324e451502e04ceefcf61a1a946640e"} Oct 05 07:08:48 crc kubenswrapper[4935]: I1005 07:08:48.217234 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-82msx" event={"ID":"171c85a1-08f1-481c-85fd-e60317147ad2","Type":"ContainerStarted","Data":"4824bf6b889b63d7b9ac051217f0534cab393aa1c63b4f31515e5e6e0024737a"} Oct 05 07:08:48 crc kubenswrapper[4935]: I1005 07:08:48.240344 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-82msx" podStartSLOduration=1.240297461 podStartE2EDuration="1.240297461s" podCreationTimestamp="2025-10-05 07:08:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:08:48.233938933 +0000 UTC m=+962.116565483" watchObservedRunningTime="2025-10-05 07:08:48.240297461 +0000 UTC m=+962.122923961" Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.528848 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fpvbl"] Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.530199 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fpvbl" Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.540446 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-f6vcs" Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.540780 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.541068 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.559186 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fpvbl"] Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.561134 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmfqq\" (UniqueName: \"kubernetes.io/projected/499206fd-7b76-4b79-acd6-909ce56affe8-kube-api-access-dmfqq\") pod \"openstack-operator-index-fpvbl\" (UID: \"499206fd-7b76-4b79-acd6-909ce56affe8\") " pod="openstack-operators/openstack-operator-index-fpvbl" Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.662002 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmfqq\" (UniqueName: \"kubernetes.io/projected/499206fd-7b76-4b79-acd6-909ce56affe8-kube-api-access-dmfqq\") pod \"openstack-operator-index-fpvbl\" (UID: \"499206fd-7b76-4b79-acd6-909ce56affe8\") " pod="openstack-operators/openstack-operator-index-fpvbl" Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.680048 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmfqq\" (UniqueName: \"kubernetes.io/projected/499206fd-7b76-4b79-acd6-909ce56affe8-kube-api-access-dmfqq\") pod \"openstack-operator-index-fpvbl\" (UID: \"499206fd-7b76-4b79-acd6-909ce56affe8\") " pod="openstack-operators/openstack-operator-index-fpvbl" Oct 05 07:08:52 crc kubenswrapper[4935]: I1005 07:08:52.861677 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fpvbl" Oct 05 07:08:53 crc kubenswrapper[4935]: I1005 07:08:53.301388 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fpvbl"] Oct 05 07:08:54 crc kubenswrapper[4935]: I1005 07:08:54.255080 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fpvbl" event={"ID":"499206fd-7b76-4b79-acd6-909ce56affe8","Type":"ContainerStarted","Data":"6b229bded29ee6e908f09f9ac1fd243d8f13514d082758c3083d252bb1a5c2c1"} Oct 05 07:08:54 crc kubenswrapper[4935]: I1005 07:08:54.255447 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fpvbl" event={"ID":"499206fd-7b76-4b79-acd6-909ce56affe8","Type":"ContainerStarted","Data":"248e2a22e21cd87e44f1dbf3122d44cf9471030a39bb8718f142b8c091040dbf"} Oct 05 07:08:54 crc kubenswrapper[4935]: I1005 07:08:54.269932 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fpvbl" podStartSLOduration=1.513331203 podStartE2EDuration="2.269915463s" podCreationTimestamp="2025-10-05 07:08:52 +0000 UTC" firstStartedPulling="2025-10-05 07:08:53.306463422 +0000 UTC m=+967.189089902" lastFinishedPulling="2025-10-05 07:08:54.063047702 +0000 UTC m=+967.945674162" observedRunningTime="2025-10-05 07:08:54.268629129 +0000 UTC m=+968.151255609" watchObservedRunningTime="2025-10-05 07:08:54.269915463 +0000 UTC m=+968.152541923" Oct 05 07:08:56 crc kubenswrapper[4935]: I1005 07:08:56.300037 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fpvbl"] Oct 05 07:08:56 crc kubenswrapper[4935]: I1005 07:08:56.300661 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-fpvbl" podUID="499206fd-7b76-4b79-acd6-909ce56affe8" containerName="registry-server" containerID="cri-o://6b229bded29ee6e908f09f9ac1fd243d8f13514d082758c3083d252bb1a5c2c1" gracePeriod=2 Oct 05 07:08:56 crc kubenswrapper[4935]: I1005 07:08:56.910107 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xzqmb"] Oct 05 07:08:56 crc kubenswrapper[4935]: I1005 07:08:56.911724 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:08:56 crc kubenswrapper[4935]: I1005 07:08:56.929446 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hjmk\" (UniqueName: \"kubernetes.io/projected/8a6a9e34-892e-4580-8cde-73d82b0c65f2-kube-api-access-7hjmk\") pod \"openstack-operator-index-xzqmb\" (UID: \"8a6a9e34-892e-4580-8cde-73d82b0c65f2\") " pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:08:56 crc kubenswrapper[4935]: I1005 07:08:56.934853 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xzqmb"] Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.030874 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hjmk\" (UniqueName: \"kubernetes.io/projected/8a6a9e34-892e-4580-8cde-73d82b0c65f2-kube-api-access-7hjmk\") pod \"openstack-operator-index-xzqmb\" (UID: \"8a6a9e34-892e-4580-8cde-73d82b0c65f2\") " pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.067645 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hjmk\" (UniqueName: \"kubernetes.io/projected/8a6a9e34-892e-4580-8cde-73d82b0c65f2-kube-api-access-7hjmk\") pod \"openstack-operator-index-xzqmb\" (UID: \"8a6a9e34-892e-4580-8cde-73d82b0c65f2\") " pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.233466 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.280245 4935 generic.go:334] "Generic (PLEG): container finished" podID="499206fd-7b76-4b79-acd6-909ce56affe8" containerID="6b229bded29ee6e908f09f9ac1fd243d8f13514d082758c3083d252bb1a5c2c1" exitCode=0 Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.280301 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fpvbl" event={"ID":"499206fd-7b76-4b79-acd6-909ce56affe8","Type":"ContainerDied","Data":"6b229bded29ee6e908f09f9ac1fd243d8f13514d082758c3083d252bb1a5c2c1"} Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.428875 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xzqmb"] Oct 05 07:08:57 crc kubenswrapper[4935]: W1005 07:08:57.435464 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a6a9e34_892e_4580_8cde_73d82b0c65f2.slice/crio-26841b931e2134c09acfc93253ec696598f804428af35da01553d9fce2467188 WatchSource:0}: Error finding container 26841b931e2134c09acfc93253ec696598f804428af35da01553d9fce2467188: Status 404 returned error can't find the container with id 26841b931e2134c09acfc93253ec696598f804428af35da01553d9fce2467188 Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.733159 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fpvbl" Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.840674 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmfqq\" (UniqueName: \"kubernetes.io/projected/499206fd-7b76-4b79-acd6-909ce56affe8-kube-api-access-dmfqq\") pod \"499206fd-7b76-4b79-acd6-909ce56affe8\" (UID: \"499206fd-7b76-4b79-acd6-909ce56affe8\") " Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.845644 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/499206fd-7b76-4b79-acd6-909ce56affe8-kube-api-access-dmfqq" (OuterVolumeSpecName: "kube-api-access-dmfqq") pod "499206fd-7b76-4b79-acd6-909ce56affe8" (UID: "499206fd-7b76-4b79-acd6-909ce56affe8"). InnerVolumeSpecName "kube-api-access-dmfqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:57 crc kubenswrapper[4935]: I1005 07:08:57.942678 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmfqq\" (UniqueName: \"kubernetes.io/projected/499206fd-7b76-4b79-acd6-909ce56affe8-kube-api-access-dmfqq\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.288044 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xzqmb" event={"ID":"8a6a9e34-892e-4580-8cde-73d82b0c65f2","Type":"ContainerStarted","Data":"04f4a6c0b43937efda2d8ae2a3c933a1838a585274c2e13409727b5fab9dbec3"} Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.288403 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xzqmb" event={"ID":"8a6a9e34-892e-4580-8cde-73d82b0c65f2","Type":"ContainerStarted","Data":"26841b931e2134c09acfc93253ec696598f804428af35da01553d9fce2467188"} Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.289543 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fpvbl" event={"ID":"499206fd-7b76-4b79-acd6-909ce56affe8","Type":"ContainerDied","Data":"248e2a22e21cd87e44f1dbf3122d44cf9471030a39bb8718f142b8c091040dbf"} Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.289595 4935 scope.go:117] "RemoveContainer" containerID="6b229bded29ee6e908f09f9ac1fd243d8f13514d082758c3083d252bb1a5c2c1" Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.289721 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fpvbl" Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.305624 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xzqmb" podStartSLOduration=1.717151927 podStartE2EDuration="2.30559836s" podCreationTimestamp="2025-10-05 07:08:56 +0000 UTC" firstStartedPulling="2025-10-05 07:08:57.439142464 +0000 UTC m=+971.321768924" lastFinishedPulling="2025-10-05 07:08:58.027588867 +0000 UTC m=+971.910215357" observedRunningTime="2025-10-05 07:08:58.300359322 +0000 UTC m=+972.182985792" watchObservedRunningTime="2025-10-05 07:08:58.30559836 +0000 UTC m=+972.188224810" Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.329340 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fpvbl"] Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.332629 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-fpvbl"] Oct 05 07:08:58 crc kubenswrapper[4935]: I1005 07:08:58.784993 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="499206fd-7b76-4b79-acd6-909ce56affe8" path="/var/lib/kubelet/pods/499206fd-7b76-4b79-acd6-909ce56affe8/volumes" Oct 05 07:09:07 crc kubenswrapper[4935]: I1005 07:09:07.233604 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:09:07 crc kubenswrapper[4935]: I1005 07:09:07.234165 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:09:07 crc kubenswrapper[4935]: I1005 07:09:07.277260 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:09:07 crc kubenswrapper[4935]: I1005 07:09:07.373640 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xzqmb" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.331246 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm"] Oct 05 07:09:09 crc kubenswrapper[4935]: E1005 07:09:09.331567 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="499206fd-7b76-4b79-acd6-909ce56affe8" containerName="registry-server" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.331581 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="499206fd-7b76-4b79-acd6-909ce56affe8" containerName="registry-server" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.331710 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="499206fd-7b76-4b79-acd6-909ce56affe8" containerName="registry-server" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.332678 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.334522 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-rnqwb" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.340227 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm"] Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.397100 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-bundle\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.397394 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82dhj\" (UniqueName: \"kubernetes.io/projected/fa8536f3-2840-4264-9e20-df1499d52611-kube-api-access-82dhj\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.397525 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-util\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.499152 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-bundle\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.499204 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82dhj\" (UniqueName: \"kubernetes.io/projected/fa8536f3-2840-4264-9e20-df1499d52611-kube-api-access-82dhj\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.499245 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-util\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.499748 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-util\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.500061 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-bundle\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.515247 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82dhj\" (UniqueName: \"kubernetes.io/projected/fa8536f3-2840-4264-9e20-df1499d52611-kube-api-access-82dhj\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:09 crc kubenswrapper[4935]: I1005 07:09:09.647123 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:10 crc kubenswrapper[4935]: I1005 07:09:10.046237 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm"] Oct 05 07:09:10 crc kubenswrapper[4935]: I1005 07:09:10.372131 4935 generic.go:334] "Generic (PLEG): container finished" podID="fa8536f3-2840-4264-9e20-df1499d52611" containerID="54374070cfa074e6eb0e07148441fdb1b22fb6e187f3d2667883f422611e2491" exitCode=0 Oct 05 07:09:10 crc kubenswrapper[4935]: I1005 07:09:10.372180 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" event={"ID":"fa8536f3-2840-4264-9e20-df1499d52611","Type":"ContainerDied","Data":"54374070cfa074e6eb0e07148441fdb1b22fb6e187f3d2667883f422611e2491"} Oct 05 07:09:10 crc kubenswrapper[4935]: I1005 07:09:10.372207 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" event={"ID":"fa8536f3-2840-4264-9e20-df1499d52611","Type":"ContainerStarted","Data":"15899ced321b9da63c3d106dd4bf00479cf5ce27b727250e84a012d007449017"} Oct 05 07:09:11 crc kubenswrapper[4935]: I1005 07:09:11.381138 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" event={"ID":"fa8536f3-2840-4264-9e20-df1499d52611","Type":"ContainerStarted","Data":"415c2f7be94a95283fe9e5a7160f0a031d9180f5e80e5bf6b1042410dd5240bc"} Oct 05 07:09:12 crc kubenswrapper[4935]: I1005 07:09:12.400516 4935 generic.go:334] "Generic (PLEG): container finished" podID="fa8536f3-2840-4264-9e20-df1499d52611" containerID="415c2f7be94a95283fe9e5a7160f0a031d9180f5e80e5bf6b1042410dd5240bc" exitCode=0 Oct 05 07:09:12 crc kubenswrapper[4935]: I1005 07:09:12.400565 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" event={"ID":"fa8536f3-2840-4264-9e20-df1499d52611","Type":"ContainerDied","Data":"415c2f7be94a95283fe9e5a7160f0a031d9180f5e80e5bf6b1042410dd5240bc"} Oct 05 07:09:13 crc kubenswrapper[4935]: I1005 07:09:13.409714 4935 generic.go:334] "Generic (PLEG): container finished" podID="fa8536f3-2840-4264-9e20-df1499d52611" containerID="543980f4a5e49489a5c80d08808e0e8e18aeccf301a09f1bc09f1a56e7bfbb24" exitCode=0 Oct 05 07:09:13 crc kubenswrapper[4935]: I1005 07:09:13.409778 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" event={"ID":"fa8536f3-2840-4264-9e20-df1499d52611","Type":"ContainerDied","Data":"543980f4a5e49489a5c80d08808e0e8e18aeccf301a09f1bc09f1a56e7bfbb24"} Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.726509 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.883882 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82dhj\" (UniqueName: \"kubernetes.io/projected/fa8536f3-2840-4264-9e20-df1499d52611-kube-api-access-82dhj\") pod \"fa8536f3-2840-4264-9e20-df1499d52611\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.884031 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-bundle\") pod \"fa8536f3-2840-4264-9e20-df1499d52611\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.884199 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-util\") pod \"fa8536f3-2840-4264-9e20-df1499d52611\" (UID: \"fa8536f3-2840-4264-9e20-df1499d52611\") " Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.885464 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-bundle" (OuterVolumeSpecName: "bundle") pod "fa8536f3-2840-4264-9e20-df1499d52611" (UID: "fa8536f3-2840-4264-9e20-df1499d52611"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.893211 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa8536f3-2840-4264-9e20-df1499d52611-kube-api-access-82dhj" (OuterVolumeSpecName: "kube-api-access-82dhj") pod "fa8536f3-2840-4264-9e20-df1499d52611" (UID: "fa8536f3-2840-4264-9e20-df1499d52611"). InnerVolumeSpecName "kube-api-access-82dhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.901333 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-util" (OuterVolumeSpecName: "util") pod "fa8536f3-2840-4264-9e20-df1499d52611" (UID: "fa8536f3-2840-4264-9e20-df1499d52611"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.986154 4935 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-util\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.986219 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82dhj\" (UniqueName: \"kubernetes.io/projected/fa8536f3-2840-4264-9e20-df1499d52611-kube-api-access-82dhj\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:14 crc kubenswrapper[4935]: I1005 07:09:14.986263 4935 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa8536f3-2840-4264-9e20-df1499d52611-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:15 crc kubenswrapper[4935]: I1005 07:09:15.429249 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" event={"ID":"fa8536f3-2840-4264-9e20-df1499d52611","Type":"ContainerDied","Data":"15899ced321b9da63c3d106dd4bf00479cf5ce27b727250e84a012d007449017"} Oct 05 07:09:15 crc kubenswrapper[4935]: I1005 07:09:15.429289 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15899ced321b9da63c3d106dd4bf00479cf5ce27b727250e84a012d007449017" Oct 05 07:09:15 crc kubenswrapper[4935]: I1005 07:09:15.429389 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.124498 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h"] Oct 05 07:09:22 crc kubenswrapper[4935]: E1005 07:09:22.125230 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8536f3-2840-4264-9e20-df1499d52611" containerName="pull" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.125243 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8536f3-2840-4264-9e20-df1499d52611" containerName="pull" Oct 05 07:09:22 crc kubenswrapper[4935]: E1005 07:09:22.125269 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8536f3-2840-4264-9e20-df1499d52611" containerName="util" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.125274 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8536f3-2840-4264-9e20-df1499d52611" containerName="util" Oct 05 07:09:22 crc kubenswrapper[4935]: E1005 07:09:22.125281 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8536f3-2840-4264-9e20-df1499d52611" containerName="extract" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.125287 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8536f3-2840-4264-9e20-df1499d52611" containerName="extract" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.125429 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa8536f3-2840-4264-9e20-df1499d52611" containerName="extract" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.126232 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.129104 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-sv4g4" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.167397 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h"] Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.283886 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpbpw\" (UniqueName: \"kubernetes.io/projected/15009458-5da1-43c5-8e7d-9341a8794532-kube-api-access-jpbpw\") pod \"openstack-operator-controller-operator-677d5bb784-9vt2h\" (UID: \"15009458-5da1-43c5-8e7d-9341a8794532\") " pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.384771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpbpw\" (UniqueName: \"kubernetes.io/projected/15009458-5da1-43c5-8e7d-9341a8794532-kube-api-access-jpbpw\") pod \"openstack-operator-controller-operator-677d5bb784-9vt2h\" (UID: \"15009458-5da1-43c5-8e7d-9341a8794532\") " pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.410972 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpbpw\" (UniqueName: \"kubernetes.io/projected/15009458-5da1-43c5-8e7d-9341a8794532-kube-api-access-jpbpw\") pod \"openstack-operator-controller-operator-677d5bb784-9vt2h\" (UID: \"15009458-5da1-43c5-8e7d-9341a8794532\") " pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.441543 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" Oct 05 07:09:22 crc kubenswrapper[4935]: I1005 07:09:22.915259 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h"] Oct 05 07:09:23 crc kubenswrapper[4935]: I1005 07:09:23.482219 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" event={"ID":"15009458-5da1-43c5-8e7d-9341a8794532","Type":"ContainerStarted","Data":"f7a272206ebd9766b7def218c0527a2ef0d2976b27b6f82d18e0a143fd690943"} Oct 05 07:09:27 crc kubenswrapper[4935]: I1005 07:09:27.508222 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" event={"ID":"15009458-5da1-43c5-8e7d-9341a8794532","Type":"ContainerStarted","Data":"f10519dd46ba9c27e74de84d54d15200844c802c440bfaf75ff447b5afcee07d"} Oct 05 07:09:29 crc kubenswrapper[4935]: I1005 07:09:29.533284 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" event={"ID":"15009458-5da1-43c5-8e7d-9341a8794532","Type":"ContainerStarted","Data":"6d94a9ccb26cb624897c0af6b76ebd59d61f955592c1c3da4670aa7fe9f1899e"} Oct 05 07:09:29 crc kubenswrapper[4935]: I1005 07:09:29.534547 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" Oct 05 07:09:29 crc kubenswrapper[4935]: I1005 07:09:29.563201 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" podStartSLOduration=1.438040278 podStartE2EDuration="7.563186584s" podCreationTimestamp="2025-10-05 07:09:22 +0000 UTC" firstStartedPulling="2025-10-05 07:09:22.931122422 +0000 UTC m=+996.813748872" lastFinishedPulling="2025-10-05 07:09:29.056268718 +0000 UTC m=+1002.938895178" observedRunningTime="2025-10-05 07:09:29.560863163 +0000 UTC m=+1003.443489623" watchObservedRunningTime="2025-10-05 07:09:29.563186584 +0000 UTC m=+1003.445813044" Oct 05 07:09:32 crc kubenswrapper[4935]: I1005 07:09:32.445740 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-9vt2h" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.740559 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.742432 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.744248 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-rpv5m" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.745692 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.746733 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.748096 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-87pbd" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.756548 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.765077 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.766242 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.770625 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-lnt5n" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.784341 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-55r29"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.788584 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.816386 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.817230 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-wnzbb" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.835360 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.835859 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz4xc\" (UniqueName: \"kubernetes.io/projected/1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e-kube-api-access-gz4xc\") pod \"glance-operator-controller-manager-698456cdc6-55r29\" (UID: \"1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.836045 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxxmj\" (UniqueName: \"kubernetes.io/projected/96b524ed-4a5d-4c95-b076-c7f68fab13a2-kube-api-access-rxxmj\") pod \"designate-operator-controller-manager-58d86cd59d-ntjg7\" (UID: \"96b524ed-4a5d-4c95-b076-c7f68fab13a2\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.836165 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnxkd\" (UniqueName: \"kubernetes.io/projected/1ef01eb4-46b6-4859-a53f-1b9b58f9f408-kube-api-access-pnxkd\") pod \"cinder-operator-controller-manager-84bd8f6848-nrs2t\" (UID: \"1ef01eb4-46b6-4859-a53f-1b9b58f9f408\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.836319 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjjjv\" (UniqueName: \"kubernetes.io/projected/451246ff-f04c-4b5b-a956-41592f58ddb9-kube-api-access-jjjjv\") pod \"barbican-operator-controller-manager-5b974f6766-6nxjj\" (UID: \"451246ff-f04c-4b5b-a956-41592f58ddb9\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.846034 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-55r29"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.864142 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.865124 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.871373 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-fjlcw" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.872610 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.884415 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.890943 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.891299 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-228pt" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.901714 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.903881 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.911566 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.911794 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.912789 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.915496 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-58lg9" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.915682 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-k5rgw" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.915792 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.915880 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.916924 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.920256 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hrkrk" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.923713 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.939169 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.940977 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z4wl\" (UniqueName: \"kubernetes.io/projected/dd907a5c-8a2c-459a-b4e0-2f1341fe956b-kube-api-access-9z4wl\") pod \"keystone-operator-controller-manager-57c9cdcf57-22j5g\" (UID: \"dd907a5c-8a2c-459a-b4e0-2f1341fe956b\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941027 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-b285k\" (UID: \"6c0af6f9-3f50-4620-94a9-9c9262d73fd9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941057 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq8wn\" (UniqueName: \"kubernetes.io/projected/eb1aa07b-d43b-4ad3-803c-042de10feab8-kube-api-access-pq8wn\") pod \"ironic-operator-controller-manager-6f5894c49f-g84c7\" (UID: \"eb1aa07b-d43b-4ad3-803c-042de10feab8\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941079 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnsg2\" (UniqueName: \"kubernetes.io/projected/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-kube-api-access-mnsg2\") pod \"infra-operator-controller-manager-84788b6bc5-b285k\" (UID: \"6c0af6f9-3f50-4620-94a9-9c9262d73fd9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941106 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxdsd\" (UniqueName: \"kubernetes.io/projected/c7430e4b-67b4-45b5-8444-1ee6cf758c5b-kube-api-access-cxdsd\") pod \"heat-operator-controller-manager-5c497dbdb-s8gzx\" (UID: \"c7430e4b-67b4-45b5-8444-1ee6cf758c5b\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941149 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjjjv\" (UniqueName: \"kubernetes.io/projected/451246ff-f04c-4b5b-a956-41592f58ddb9-kube-api-access-jjjjv\") pod \"barbican-operator-controller-manager-5b974f6766-6nxjj\" (UID: \"451246ff-f04c-4b5b-a956-41592f58ddb9\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941221 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz4xc\" (UniqueName: \"kubernetes.io/projected/1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e-kube-api-access-gz4xc\") pod \"glance-operator-controller-manager-698456cdc6-55r29\" (UID: \"1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941267 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dqt9\" (UniqueName: \"kubernetes.io/projected/317d3174-23c3-4af7-b289-26844e8d3f12-kube-api-access-9dqt9\") pod \"horizon-operator-controller-manager-6675647785-zk5fs\" (UID: \"317d3174-23c3-4af7-b289-26844e8d3f12\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941313 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxxmj\" (UniqueName: \"kubernetes.io/projected/96b524ed-4a5d-4c95-b076-c7f68fab13a2-kube-api-access-rxxmj\") pod \"designate-operator-controller-manager-58d86cd59d-ntjg7\" (UID: \"96b524ed-4a5d-4c95-b076-c7f68fab13a2\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.941347 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnxkd\" (UniqueName: \"kubernetes.io/projected/1ef01eb4-46b6-4859-a53f-1b9b58f9f408-kube-api-access-pnxkd\") pod \"cinder-operator-controller-manager-84bd8f6848-nrs2t\" (UID: \"1ef01eb4-46b6-4859-a53f-1b9b58f9f408\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.942464 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.952392 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.957711 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-jtc6k" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.969099 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.979679 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnxkd\" (UniqueName: \"kubernetes.io/projected/1ef01eb4-46b6-4859-a53f-1b9b58f9f408-kube-api-access-pnxkd\") pod \"cinder-operator-controller-manager-84bd8f6848-nrs2t\" (UID: \"1ef01eb4-46b6-4859-a53f-1b9b58f9f408\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.991855 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz4xc\" (UniqueName: \"kubernetes.io/projected/1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e-kube-api-access-gz4xc\") pod \"glance-operator-controller-manager-698456cdc6-55r29\" (UID: \"1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.992068 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.992126 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.992836 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxxmj\" (UniqueName: \"kubernetes.io/projected/96b524ed-4a5d-4c95-b076-c7f68fab13a2-kube-api-access-rxxmj\") pod \"designate-operator-controller-manager-58d86cd59d-ntjg7\" (UID: \"96b524ed-4a5d-4c95-b076-c7f68fab13a2\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.993122 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.993993 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65"] Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.995223 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.998459 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-wb4cg" Oct 05 07:09:47 crc kubenswrapper[4935]: I1005 07:09:47.999158 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-hdgmc" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.008084 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjjjv\" (UniqueName: \"kubernetes.io/projected/451246ff-f04c-4b5b-a956-41592f58ddb9-kube-api-access-jjjjv\") pod \"barbican-operator-controller-manager-5b974f6766-6nxjj\" (UID: \"451246ff-f04c-4b5b-a956-41592f58ddb9\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.013991 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.014960 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.021012 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.023643 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-l5d6r" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.025107 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.036643 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.038084 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.040714 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.041719 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-brfdm" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.042834 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxdsd\" (UniqueName: \"kubernetes.io/projected/c7430e4b-67b4-45b5-8444-1ee6cf758c5b-kube-api-access-cxdsd\") pod \"heat-operator-controller-manager-5c497dbdb-s8gzx\" (UID: \"c7430e4b-67b4-45b5-8444-1ee6cf758c5b\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.042916 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nnx6\" (UniqueName: \"kubernetes.io/projected/3078f420-c252-4650-ab0a-17c92ab8d1c6-kube-api-access-4nnx6\") pod \"manila-operator-controller-manager-7cb48dbc-lhzb8\" (UID: \"3078f420-c252-4650-ab0a-17c92ab8d1c6\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.042953 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj54q\" (UniqueName: \"kubernetes.io/projected/1fb98234-7f3b-491f-871a-106ce0f2af2a-kube-api-access-tj54q\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-llst6\" (UID: \"1fb98234-7f3b-491f-871a-106ce0f2af2a\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.042972 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrcks\" (UniqueName: \"kubernetes.io/projected/39f750d8-d3c5-4164-ab98-12c031e752ab-kube-api-access-rrcks\") pod \"octavia-operator-controller-manager-69f59f9d8-d2l5q\" (UID: \"39f750d8-d3c5-4164-ab98-12c031e752ab\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.043030 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dqt9\" (UniqueName: \"kubernetes.io/projected/317d3174-23c3-4af7-b289-26844e8d3f12-kube-api-access-9dqt9\") pod \"horizon-operator-controller-manager-6675647785-zk5fs\" (UID: \"317d3174-23c3-4af7-b289-26844e8d3f12\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.043050 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcqg9\" (UniqueName: \"kubernetes.io/projected/1201b72a-0267-4e34-bdf5-6ddb0f9299dd-kube-api-access-jcqg9\") pod \"nova-operator-controller-manager-6c9b57c67-l9bf8\" (UID: \"1201b72a-0267-4e34-bdf5-6ddb0f9299dd\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.043070 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhcqb\" (UniqueName: \"kubernetes.io/projected/5b24bc52-99be-499c-90ac-91114405b8f6-kube-api-access-lhcqb\") pod \"neutron-operator-controller-manager-69b956fbf6-55k65\" (UID: \"5b24bc52-99be-499c-90ac-91114405b8f6\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.043110 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z4wl\" (UniqueName: \"kubernetes.io/projected/dd907a5c-8a2c-459a-b4e0-2f1341fe956b-kube-api-access-9z4wl\") pod \"keystone-operator-controller-manager-57c9cdcf57-22j5g\" (UID: \"dd907a5c-8a2c-459a-b4e0-2f1341fe956b\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.043130 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-b285k\" (UID: \"6c0af6f9-3f50-4620-94a9-9c9262d73fd9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.043147 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq8wn\" (UniqueName: \"kubernetes.io/projected/eb1aa07b-d43b-4ad3-803c-042de10feab8-kube-api-access-pq8wn\") pod \"ironic-operator-controller-manager-6f5894c49f-g84c7\" (UID: \"eb1aa07b-d43b-4ad3-803c-042de10feab8\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.043162 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnsg2\" (UniqueName: \"kubernetes.io/projected/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-kube-api-access-mnsg2\") pod \"infra-operator-controller-manager-84788b6bc5-b285k\" (UID: \"6c0af6f9-3f50-4620-94a9-9c9262d73fd9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:48 crc kubenswrapper[4935]: E1005 07:09:48.044153 4935 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 05 07:09:48 crc kubenswrapper[4935]: E1005 07:09:48.044281 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-cert podName:6c0af6f9-3f50-4620-94a9-9c9262d73fd9 nodeName:}" failed. No retries permitted until 2025-10-05 07:09:48.544260364 +0000 UTC m=+1022.426886834 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-cert") pod "infra-operator-controller-manager-84788b6bc5-b285k" (UID: "6c0af6f9-3f50-4620-94a9-9c9262d73fd9") : secret "infra-operator-webhook-server-cert" not found Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.053491 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.059346 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.065841 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.066427 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-g2828"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.077583 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.080342 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dqt9\" (UniqueName: \"kubernetes.io/projected/317d3174-23c3-4af7-b289-26844e8d3f12-kube-api-access-9dqt9\") pod \"horizon-operator-controller-manager-6675647785-zk5fs\" (UID: \"317d3174-23c3-4af7-b289-26844e8d3f12\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.080625 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-wqkww" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.089435 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z4wl\" (UniqueName: \"kubernetes.io/projected/dd907a5c-8a2c-459a-b4e0-2f1341fe956b-kube-api-access-9z4wl\") pod \"keystone-operator-controller-manager-57c9cdcf57-22j5g\" (UID: \"dd907a5c-8a2c-459a-b4e0-2f1341fe956b\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.089856 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnsg2\" (UniqueName: \"kubernetes.io/projected/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-kube-api-access-mnsg2\") pod \"infra-operator-controller-manager-84788b6bc5-b285k\" (UID: \"6c0af6f9-3f50-4620-94a9-9c9262d73fd9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.090714 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq8wn\" (UniqueName: \"kubernetes.io/projected/eb1aa07b-d43b-4ad3-803c-042de10feab8-kube-api-access-pq8wn\") pod \"ironic-operator-controller-manager-6f5894c49f-g84c7\" (UID: \"eb1aa07b-d43b-4ad3-803c-042de10feab8\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.092330 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.096602 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxdsd\" (UniqueName: \"kubernetes.io/projected/c7430e4b-67b4-45b5-8444-1ee6cf758c5b-kube-api-access-cxdsd\") pod \"heat-operator-controller-manager-5c497dbdb-s8gzx\" (UID: \"c7430e4b-67b4-45b5-8444-1ee6cf758c5b\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.141623 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.142767 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.148199 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.149171 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-7gcc2" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.150020 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nnx6\" (UniqueName: \"kubernetes.io/projected/3078f420-c252-4650-ab0a-17c92ab8d1c6-kube-api-access-4nnx6\") pod \"manila-operator-controller-manager-7cb48dbc-lhzb8\" (UID: \"3078f420-c252-4650-ab0a-17c92ab8d1c6\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.150107 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz\" (UID: \"1a18b889-577c-4bb0-9d34-b2e560318144\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.150183 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj54q\" (UniqueName: \"kubernetes.io/projected/1fb98234-7f3b-491f-871a-106ce0f2af2a-kube-api-access-tj54q\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-llst6\" (UID: \"1fb98234-7f3b-491f-871a-106ce0f2af2a\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.150219 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrcks\" (UniqueName: \"kubernetes.io/projected/39f750d8-d3c5-4164-ab98-12c031e752ab-kube-api-access-rrcks\") pod \"octavia-operator-controller-manager-69f59f9d8-d2l5q\" (UID: \"39f750d8-d3c5-4164-ab98-12c031e752ab\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.150253 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcqg9\" (UniqueName: \"kubernetes.io/projected/1201b72a-0267-4e34-bdf5-6ddb0f9299dd-kube-api-access-jcqg9\") pod \"nova-operator-controller-manager-6c9b57c67-l9bf8\" (UID: \"1201b72a-0267-4e34-bdf5-6ddb0f9299dd\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.150289 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhcqb\" (UniqueName: \"kubernetes.io/projected/5b24bc52-99be-499c-90ac-91114405b8f6-kube-api-access-lhcqb\") pod \"neutron-operator-controller-manager-69b956fbf6-55k65\" (UID: \"5b24bc52-99be-499c-90ac-91114405b8f6\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.150329 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq6bn\" (UniqueName: \"kubernetes.io/projected/1a18b889-577c-4bb0-9d34-b2e560318144-kube-api-access-vq6bn\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz\" (UID: \"1a18b889-577c-4bb0-9d34-b2e560318144\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.150458 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdd74\" (UniqueName: \"kubernetes.io/projected/57067c6a-4af7-4b0f-93cf-ac544f00a239-kube-api-access-mdd74\") pod \"ovn-operator-controller-manager-c968bb45-g2828\" (UID: \"57067c6a-4af7-4b0f-93cf-ac544f00a239\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.155116 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-g2828"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.157693 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.186822 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj54q\" (UniqueName: \"kubernetes.io/projected/1fb98234-7f3b-491f-871a-106ce0f2af2a-kube-api-access-tj54q\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-llst6\" (UID: \"1fb98234-7f3b-491f-871a-106ce0f2af2a\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.190030 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.196460 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nnx6\" (UniqueName: \"kubernetes.io/projected/3078f420-c252-4650-ab0a-17c92ab8d1c6-kube-api-access-4nnx6\") pod \"manila-operator-controller-manager-7cb48dbc-lhzb8\" (UID: \"3078f420-c252-4650-ab0a-17c92ab8d1c6\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.202783 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.202798 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.205262 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-hvnf2" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.207999 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcqg9\" (UniqueName: \"kubernetes.io/projected/1201b72a-0267-4e34-bdf5-6ddb0f9299dd-kube-api-access-jcqg9\") pod \"nova-operator-controller-manager-6c9b57c67-l9bf8\" (UID: \"1201b72a-0267-4e34-bdf5-6ddb0f9299dd\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.209254 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhcqb\" (UniqueName: \"kubernetes.io/projected/5b24bc52-99be-499c-90ac-91114405b8f6-kube-api-access-lhcqb\") pod \"neutron-operator-controller-manager-69b956fbf6-55k65\" (UID: \"5b24bc52-99be-499c-90ac-91114405b8f6\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.212820 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.218396 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrcks\" (UniqueName: \"kubernetes.io/projected/39f750d8-d3c5-4164-ab98-12c031e752ab-kube-api-access-rrcks\") pod \"octavia-operator-controller-manager-69f59f9d8-d2l5q\" (UID: \"39f750d8-d3c5-4164-ab98-12c031e752ab\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.218836 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.239621 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.239940 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.254075 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz\" (UID: \"1a18b889-577c-4bb0-9d34-b2e560318144\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.254138 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq6bn\" (UniqueName: \"kubernetes.io/projected/1a18b889-577c-4bb0-9d34-b2e560318144-kube-api-access-vq6bn\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz\" (UID: \"1a18b889-577c-4bb0-9d34-b2e560318144\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.254197 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdd74\" (UniqueName: \"kubernetes.io/projected/57067c6a-4af7-4b0f-93cf-ac544f00a239-kube-api-access-mdd74\") pod \"ovn-operator-controller-manager-c968bb45-g2828\" (UID: \"57067c6a-4af7-4b0f-93cf-ac544f00a239\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" Oct 05 07:09:48 crc kubenswrapper[4935]: E1005 07:09:48.254572 4935 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 07:09:48 crc kubenswrapper[4935]: E1005 07:09:48.254864 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert podName:1a18b889-577c-4bb0-9d34-b2e560318144 nodeName:}" failed. No retries permitted until 2025-10-05 07:09:48.754831339 +0000 UTC m=+1022.637457799 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert") pod "openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" (UID: "1a18b889-577c-4bb0-9d34-b2e560318144") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.268845 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.320785 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdd74\" (UniqueName: \"kubernetes.io/projected/57067c6a-4af7-4b0f-93cf-ac544f00a239-kube-api-access-mdd74\") pod \"ovn-operator-controller-manager-c968bb45-g2828\" (UID: \"57067c6a-4af7-4b0f-93cf-ac544f00a239\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.328212 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.341559 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.343723 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq6bn\" (UniqueName: \"kubernetes.io/projected/1a18b889-577c-4bb0-9d34-b2e560318144-kube-api-access-vq6bn\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz\" (UID: \"1a18b889-577c-4bb0-9d34-b2e560318144\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.352169 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.357229 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n58x9\" (UniqueName: \"kubernetes.io/projected/8ca55603-341c-4a88-aaec-34ce00cec8fb-kube-api-access-n58x9\") pod \"swift-operator-controller-manager-76d5577b-8xv6f\" (UID: \"8ca55603-341c-4a88-aaec-34ce00cec8fb\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.357386 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqm62\" (UniqueName: \"kubernetes.io/projected/5ecfad04-5c99-47dd-b1f5-782b0aea4061-kube-api-access-gqm62\") pod \"placement-operator-controller-manager-66f6d6849b-zcr8x\" (UID: \"5ecfad04-5c99-47dd-b1f5-782b0aea4061\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.363249 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.364292 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.369241 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-hwc9q" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.369282 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.384206 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.461500 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqm62\" (UniqueName: \"kubernetes.io/projected/5ecfad04-5c99-47dd-b1f5-782b0aea4061-kube-api-access-gqm62\") pod \"placement-operator-controller-manager-66f6d6849b-zcr8x\" (UID: \"5ecfad04-5c99-47dd-b1f5-782b0aea4061\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.461595 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n58x9\" (UniqueName: \"kubernetes.io/projected/8ca55603-341c-4a88-aaec-34ce00cec8fb-kube-api-access-n58x9\") pod \"swift-operator-controller-manager-76d5577b-8xv6f\" (UID: \"8ca55603-341c-4a88-aaec-34ce00cec8fb\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.487269 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.487568 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n58x9\" (UniqueName: \"kubernetes.io/projected/8ca55603-341c-4a88-aaec-34ce00cec8fb-kube-api-access-n58x9\") pod \"swift-operator-controller-manager-76d5577b-8xv6f\" (UID: \"8ca55603-341c-4a88-aaec-34ce00cec8fb\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.487706 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.489166 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.492660 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-8nmdp" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.520503 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.540697 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.542163 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.553318 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-dtpz6" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.561991 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.568873 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.569749 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-b285k\" (UID: \"6c0af6f9-3f50-4620-94a9-9c9262d73fd9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.591336 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c0af6f9-3f50-4620-94a9-9c9262d73fd9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-b285k\" (UID: \"6c0af6f9-3f50-4620-94a9-9c9262d73fd9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.633012 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.634179 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.639149 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-6mfxd" Oct 05 07:09:48 crc kubenswrapper[4935]: W1005 07:09:48.644616 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ef01eb4_46b6_4859_a53f_1b9b58f9f408.slice/crio-9d629fa8aa0e5bcd3c4be3a658bfbe3bbcd3e12e07dc75bad8dfc46cca2b7f0b WatchSource:0}: Error finding container 9d629fa8aa0e5bcd3c4be3a658bfbe3bbcd3e12e07dc75bad8dfc46cca2b7f0b: Status 404 returned error can't find the container with id 9d629fa8aa0e5bcd3c4be3a658bfbe3bbcd3e12e07dc75bad8dfc46cca2b7f0b Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.665388 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.673602 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2f77\" (UniqueName: \"kubernetes.io/projected/8a31c4d9-3e39-4c59-ac21-f202677474ee-kube-api-access-s2f77\") pod \"test-operator-controller-manager-6bb6dcddc-hg5ns\" (UID: \"8a31c4d9-3e39-4c59-ac21-f202677474ee\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.673655 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrxpk\" (UniqueName: \"kubernetes.io/projected/446adb3e-02c1-4aa1-af8b-a038b43a3a61-kube-api-access-vrxpk\") pod \"telemetry-operator-controller-manager-f589c7597-xkph9\" (UID: \"446adb3e-02c1-4aa1-af8b-a038b43a3a61\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.716466 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqm62\" (UniqueName: \"kubernetes.io/projected/5ecfad04-5c99-47dd-b1f5-782b0aea4061-kube-api-access-gqm62\") pod \"placement-operator-controller-manager-66f6d6849b-zcr8x\" (UID: \"5ecfad04-5c99-47dd-b1f5-782b0aea4061\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.723294 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.724773 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.726754 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.729998 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-9h4z6" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.761443 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.767165 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.774961 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2f77\" (UniqueName: \"kubernetes.io/projected/8a31c4d9-3e39-4c59-ac21-f202677474ee-kube-api-access-s2f77\") pod \"test-operator-controller-manager-6bb6dcddc-hg5ns\" (UID: \"8a31c4d9-3e39-4c59-ac21-f202677474ee\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.775015 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz\" (UID: \"1a18b889-577c-4bb0-9d34-b2e560318144\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.775056 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrxpk\" (UniqueName: \"kubernetes.io/projected/446adb3e-02c1-4aa1-af8b-a038b43a3a61-kube-api-access-vrxpk\") pod \"telemetry-operator-controller-manager-f589c7597-xkph9\" (UID: \"446adb3e-02c1-4aa1-af8b-a038b43a3a61\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.775136 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgz2l\" (UniqueName: \"kubernetes.io/projected/57af99f4-85fe-426a-a608-6c46d72cc16f-kube-api-access-kgz2l\") pod \"watcher-operator-controller-manager-5d98cc5575-rtvd8\" (UID: \"57af99f4-85fe-426a-a608-6c46d72cc16f\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" Oct 05 07:09:48 crc kubenswrapper[4935]: E1005 07:09:48.775605 4935 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 07:09:48 crc kubenswrapper[4935]: E1005 07:09:48.775652 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert podName:1a18b889-577c-4bb0-9d34-b2e560318144 nodeName:}" failed. No retries permitted until 2025-10-05 07:09:49.775636674 +0000 UTC m=+1023.658263134 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert") pod "openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" (UID: "1a18b889-577c-4bb0-9d34-b2e560318144") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.816705 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2f77\" (UniqueName: \"kubernetes.io/projected/8a31c4d9-3e39-4c59-ac21-f202677474ee-kube-api-access-s2f77\") pod \"test-operator-controller-manager-6bb6dcddc-hg5ns\" (UID: \"8a31c4d9-3e39-4c59-ac21-f202677474ee\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.817259 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrxpk\" (UniqueName: \"kubernetes.io/projected/446adb3e-02c1-4aa1-af8b-a038b43a3a61-kube-api-access-vrxpk\") pod \"telemetry-operator-controller-manager-f589c7597-xkph9\" (UID: \"446adb3e-02c1-4aa1-af8b-a038b43a3a61\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.829731 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.838302 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.839233 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.839260 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.839352 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.841817 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-hf7gv" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.844961 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.875977 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgz2l\" (UniqueName: \"kubernetes.io/projected/57af99f4-85fe-426a-a608-6c46d72cc16f-kube-api-access-kgz2l\") pod \"watcher-operator-controller-manager-5d98cc5575-rtvd8\" (UID: \"57af99f4-85fe-426a-a608-6c46d72cc16f\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.876051 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcqtg\" (UniqueName: \"kubernetes.io/projected/3ca2bfab-d801-4cbf-bb97-c0eff17db807-kube-api-access-gcqtg\") pod \"openstack-operator-controller-manager-7cfc658b9-vckkj\" (UID: \"3ca2bfab-d801-4cbf-bb97-c0eff17db807\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.876096 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert\") pod \"openstack-operator-controller-manager-7cfc658b9-vckkj\" (UID: \"3ca2bfab-d801-4cbf-bb97-c0eff17db807\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.894348 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.898475 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgz2l\" (UniqueName: \"kubernetes.io/projected/57af99f4-85fe-426a-a608-6c46d72cc16f-kube-api-access-kgz2l\") pod \"watcher-operator-controller-manager-5d98cc5575-rtvd8\" (UID: \"57af99f4-85fe-426a-a608-6c46d72cc16f\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.904171 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.942130 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" Oct 05 07:09:48 crc kubenswrapper[4935]: W1005 07:09:48.942703 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod451246ff_f04c_4b5b_a956_41592f58ddb9.slice/crio-7d5ab2afb1fd941c84344c63246dcdbcf9890b6452077144eb7f890bc3a7585a WatchSource:0}: Error finding container 7d5ab2afb1fd941c84344c63246dcdbcf9890b6452077144eb7f890bc3a7585a: Status 404 returned error can't find the container with id 7d5ab2afb1fd941c84344c63246dcdbcf9890b6452077144eb7f890bc3a7585a Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.954402 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj"] Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.978354 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbdmr\" (UniqueName: \"kubernetes.io/projected/98df2e22-67cc-485b-856d-ce3e358b0bfd-kube-api-access-nbdmr\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt\" (UID: \"98df2e22-67cc-485b-856d-ce3e358b0bfd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.978459 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcqtg\" (UniqueName: \"kubernetes.io/projected/3ca2bfab-d801-4cbf-bb97-c0eff17db807-kube-api-access-gcqtg\") pod \"openstack-operator-controller-manager-7cfc658b9-vckkj\" (UID: \"3ca2bfab-d801-4cbf-bb97-c0eff17db807\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.978506 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert\") pod \"openstack-operator-controller-manager-7cfc658b9-vckkj\" (UID: \"3ca2bfab-d801-4cbf-bb97-c0eff17db807\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:48 crc kubenswrapper[4935]: E1005 07:09:48.978621 4935 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 05 07:09:48 crc kubenswrapper[4935]: E1005 07:09:48.978668 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert podName:3ca2bfab-d801-4cbf-bb97-c0eff17db807 nodeName:}" failed. No retries permitted until 2025-10-05 07:09:49.478653919 +0000 UTC m=+1023.361280369 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert") pod "openstack-operator-controller-manager-7cfc658b9-vckkj" (UID: "3ca2bfab-d801-4cbf-bb97-c0eff17db807") : secret "webhook-server-cert" not found Oct 05 07:09:48 crc kubenswrapper[4935]: I1005 07:09:48.985651 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.012282 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcqtg\" (UniqueName: \"kubernetes.io/projected/3ca2bfab-d801-4cbf-bb97-c0eff17db807-kube-api-access-gcqtg\") pod \"openstack-operator-controller-manager-7cfc658b9-vckkj\" (UID: \"3ca2bfab-d801-4cbf-bb97-c0eff17db807\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.080210 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbdmr\" (UniqueName: \"kubernetes.io/projected/98df2e22-67cc-485b-856d-ce3e358b0bfd-kube-api-access-nbdmr\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt\" (UID: \"98df2e22-67cc-485b-856d-ce3e358b0bfd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.103793 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbdmr\" (UniqueName: \"kubernetes.io/projected/98df2e22-67cc-485b-856d-ce3e358b0bfd-kube-api-access-nbdmr\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt\" (UID: \"98df2e22-67cc-485b-856d-ce3e358b0bfd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.112265 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.163635 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.309228 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-55r29"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.343152 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx"] Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.353695 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7430e4b_67b4_45b5_8444_1ee6cf758c5b.slice/crio-499756e29f3f585c6e7623ae09c7614cc7e1d962fd990632702240803068ad15 WatchSource:0}: Error finding container 499756e29f3f585c6e7623ae09c7614cc7e1d962fd990632702240803068ad15: Status 404 returned error can't find the container with id 499756e29f3f585c6e7623ae09c7614cc7e1d962fd990632702240803068ad15 Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.376706 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb1aa07b_d43b_4ad3_803c_042de10feab8.slice/crio-32a17ca80e6c0a4ca56ec4705bd5a3beb6958770c2b5f6c58183923f49476561 WatchSource:0}: Error finding container 32a17ca80e6c0a4ca56ec4705bd5a3beb6958770c2b5f6c58183923f49476561: Status 404 returned error can't find the container with id 32a17ca80e6c0a4ca56ec4705bd5a3beb6958770c2b5f6c58183923f49476561 Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.377786 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.486808 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert\") pod \"openstack-operator-controller-manager-7cfc658b9-vckkj\" (UID: \"3ca2bfab-d801-4cbf-bb97-c0eff17db807\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.487268 4935 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.487326 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert podName:3ca2bfab-d801-4cbf-bb97-c0eff17db807 nodeName:}" failed. No retries permitted until 2025-10-05 07:09:50.487307552 +0000 UTC m=+1024.369934012 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert") pod "openstack-operator-controller-manager-7cfc658b9-vckkj" (UID: "3ca2bfab-d801-4cbf-bb97-c0eff17db807") : secret "webhook-server-cert" not found Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.521265 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.537725 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.549996 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs"] Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.553872 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod317d3174_23c3_4af7_b289_26844e8d3f12.slice/crio-f61bc489dd85e7db018b220300ad2509b5dbe622c106b876d4d978190d612096 WatchSource:0}: Error finding container f61bc489dd85e7db018b220300ad2509b5dbe622c106b876d4d978190d612096: Status 404 returned error can't find the container with id f61bc489dd85e7db018b220300ad2509b5dbe622c106b876d4d978190d612096 Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.555007 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd907a5c_8a2c_459a_b4e0_2f1341fe956b.slice/crio-64343600458c33b16bdf184ed294a6a7bae7ea28429701add89dfe5557ee4af3 WatchSource:0}: Error finding container 64343600458c33b16bdf184ed294a6a7bae7ea28429701add89dfe5557ee4af3: Status 404 returned error can't find the container with id 64343600458c33b16bdf184ed294a6a7bae7ea28429701add89dfe5557ee4af3 Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.556052 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3078f420_c252_4650_ab0a_17c92ab8d1c6.slice/crio-d570d69e8e457bad2b524dfaacae7284ae386ed6c826e7eadf967707efe64aa0 WatchSource:0}: Error finding container d570d69e8e457bad2b524dfaacae7284ae386ed6c826e7eadf967707efe64aa0: Status 404 returned error can't find the container with id d570d69e8e457bad2b524dfaacae7284ae386ed6c826e7eadf967707efe64aa0 Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.597209 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.601813 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.617355 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.621075 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.627870 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-g2828"] Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.629054 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b24bc52_99be_499c_90ac_91114405b8f6.slice/crio-e2431314a7d72e5b2458bea5c3b9ae90f78b91860ee71685618cb8b0cc9436a2 WatchSource:0}: Error finding container e2431314a7d72e5b2458bea5c3b9ae90f78b91860ee71685618cb8b0cc9436a2: Status 404 returned error can't find the container with id e2431314a7d72e5b2458bea5c3b9ae90f78b91860ee71685618cb8b0cc9436a2 Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.630631 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57067c6a_4af7_4b0f_93cf_ac544f00a239.slice/crio-e91ef5ad1845e730bfe1c9ced7ec307dda947ccd0eff100c46563387207d5477 WatchSource:0}: Error finding container e91ef5ad1845e730bfe1c9ced7ec307dda947ccd0eff100c46563387207d5477: Status 404 returned error can't find the container with id e91ef5ad1845e730bfe1c9ced7ec307dda947ccd0eff100c46563387207d5477 Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.638909 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mdd74,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-c968bb45-g2828_openstack-operators(57067c6a-4af7-4b0f-93cf-ac544f00a239): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.689460 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" event={"ID":"8ca55603-341c-4a88-aaec-34ce00cec8fb","Type":"ContainerStarted","Data":"df1a60ae5641ba259267b62268ecc61772d4c18e49615e444866c41fd30ee4ed"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.691414 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" event={"ID":"57067c6a-4af7-4b0f-93cf-ac544f00a239","Type":"ContainerStarted","Data":"e91ef5ad1845e730bfe1c9ced7ec307dda947ccd0eff100c46563387207d5477"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.695002 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" event={"ID":"5b24bc52-99be-499c-90ac-91114405b8f6","Type":"ContainerStarted","Data":"e2431314a7d72e5b2458bea5c3b9ae90f78b91860ee71685618cb8b0cc9436a2"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.696453 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" event={"ID":"dd907a5c-8a2c-459a-b4e0-2f1341fe956b","Type":"ContainerStarted","Data":"64343600458c33b16bdf184ed294a6a7bae7ea28429701add89dfe5557ee4af3"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.697723 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" event={"ID":"96b524ed-4a5d-4c95-b076-c7f68fab13a2","Type":"ContainerStarted","Data":"33a79c266fb256ce38632e64f16b412bfe4c86df10ea1f23a739be2190a0b5f8"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.698829 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" event={"ID":"1ef01eb4-46b6-4859-a53f-1b9b58f9f408","Type":"ContainerStarted","Data":"9d629fa8aa0e5bcd3c4be3a658bfbe3bbcd3e12e07dc75bad8dfc46cca2b7f0b"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.700133 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" event={"ID":"c7430e4b-67b4-45b5-8444-1ee6cf758c5b","Type":"ContainerStarted","Data":"499756e29f3f585c6e7623ae09c7614cc7e1d962fd990632702240803068ad15"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.701826 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" event={"ID":"3078f420-c252-4650-ab0a-17c92ab8d1c6","Type":"ContainerStarted","Data":"d570d69e8e457bad2b524dfaacae7284ae386ed6c826e7eadf967707efe64aa0"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.704597 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" event={"ID":"1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e","Type":"ContainerStarted","Data":"4c3446f8caf7599f3f1ccedd936fac965c22349cd7343d45d0dfd6c8324bb910"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.705811 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" event={"ID":"317d3174-23c3-4af7-b289-26844e8d3f12","Type":"ContainerStarted","Data":"f61bc489dd85e7db018b220300ad2509b5dbe622c106b876d4d978190d612096"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.707398 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" event={"ID":"451246ff-f04c-4b5b-a956-41592f58ddb9","Type":"ContainerStarted","Data":"7d5ab2afb1fd941c84344c63246dcdbcf9890b6452077144eb7f890bc3a7585a"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.708835 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" event={"ID":"1201b72a-0267-4e34-bdf5-6ddb0f9299dd","Type":"ContainerStarted","Data":"2089088bbd4f0c8148e322ebb387f2802f16c88c15b758da504ad7dcce0a86db"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.709691 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" event={"ID":"eb1aa07b-d43b-4ad3-803c-042de10feab8","Type":"ContainerStarted","Data":"32a17ca80e6c0a4ca56ec4705bd5a3beb6958770c2b5f6c58183923f49476561"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.712192 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" event={"ID":"1fb98234-7f3b-491f-871a-106ce0f2af2a","Type":"ContainerStarted","Data":"d2669806662343881a8ca0416a9bec1e10268e128852a58d929647d3b9e10643"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.713658 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" event={"ID":"39f750d8-d3c5-4164-ab98-12c031e752ab","Type":"ContainerStarted","Data":"ab50fe4cfaf595152469dd48bd84230a2f5b6835fb79191d314b845e35606fd2"} Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.787783 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9"] Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.789433 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod446adb3e_02c1_4aa1_af8b_a038b43a3a61.slice/crio-d0e85cf8b0ba3fc0e0c3403670c212c22550a046a59366e1f037c95dd437a5e3 WatchSource:0}: Error finding container d0e85cf8b0ba3fc0e0c3403670c212c22550a046a59366e1f037c95dd437a5e3: Status 404 returned error can't find the container with id d0e85cf8b0ba3fc0e0c3403670c212c22550a046a59366e1f037c95dd437a5e3 Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.792692 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.794430 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz\" (UID: \"1a18b889-577c-4bb0-9d34-b2e560318144\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.795149 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vrxpk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-f589c7597-xkph9_openstack-operators(446adb3e-02c1-4aa1-af8b-a038b43a3a61): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.800798 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.801571 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a18b889-577c-4bb0-9d34-b2e560318144-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz\" (UID: \"1a18b889-577c-4bb0-9d34-b2e560318144\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.805019 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8"] Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.811744 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" podUID="57067c6a-4af7-4b0f-93cf-ac544f00a239" Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.814773 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a31c4d9_3e39_4c59_ac21_f202677474ee.slice/crio-8bbac22b0a86e4bf0cb73f07175fdb57cf9fd3cfeac0ca0a807465d23c09cbf9 WatchSource:0}: Error finding container 8bbac22b0a86e4bf0cb73f07175fdb57cf9fd3cfeac0ca0a807465d23c09cbf9: Status 404 returned error can't find the container with id 8bbac22b0a86e4bf0cb73f07175fdb57cf9fd3cfeac0ca0a807465d23c09cbf9 Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.816445 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s2f77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-6bb6dcddc-hg5ns_openstack-operators(8a31c4d9-3e39-4c59-ac21-f202677474ee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.820862 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57af99f4_85fe_426a_a608_6c46d72cc16f.slice/crio-4b46ab3563a3e941b9bf0ab00e823f922f5a2ae9dde1555f9b2c6a0ee6f1428d WatchSource:0}: Error finding container 4b46ab3563a3e941b9bf0ab00e823f922f5a2ae9dde1555f9b2c6a0ee6f1428d: Status 404 returned error can't find the container with id 4b46ab3563a3e941b9bf0ab00e823f922f5a2ae9dde1555f9b2c6a0ee6f1428d Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.821780 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.829794 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kgz2l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5d98cc5575-rtvd8_openstack-operators(57af99f4-85fe-426a-a608-6c46d72cc16f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.902683 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt"] Oct 05 07:09:49 crc kubenswrapper[4935]: I1005 07:09:49.908748 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x"] Oct 05 07:09:49 crc kubenswrapper[4935]: W1005 07:09:49.909752 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ecfad04_5c99_47dd_b1f5_782b0aea4061.slice/crio-44d5ff503efe00409762fc451ce6408628b724692d994c959d1b9161a685e4e6 WatchSource:0}: Error finding container 44d5ff503efe00409762fc451ce6408628b724692d994c959d1b9161a685e4e6: Status 404 returned error can't find the container with id 44d5ff503efe00409762fc451ce6408628b724692d994c959d1b9161a685e4e6 Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.912629 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqm62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-66f6d6849b-zcr8x_openstack-operators(5ecfad04-5c99-47dd-b1f5-782b0aea4061): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.921574 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nbdmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt_openstack-operators(98df2e22-67cc-485b-856d-ce3e358b0bfd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.923279 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" podUID="98df2e22-67cc-485b-856d-ce3e358b0bfd" Oct 05 07:09:49 crc kubenswrapper[4935]: E1005 07:09:49.945740 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" podUID="446adb3e-02c1-4aa1-af8b-a038b43a3a61" Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.101915 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" podUID="8a31c4d9-3e39-4c59-ac21-f202677474ee" Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.113712 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" podUID="57af99f4-85fe-426a-a608-6c46d72cc16f" Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.204588 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" podUID="5ecfad04-5c99-47dd-b1f5-782b0aea4061" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.342609 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz"] Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.507478 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert\") pod \"openstack-operator-controller-manager-7cfc658b9-vckkj\" (UID: \"3ca2bfab-d801-4cbf-bb97-c0eff17db807\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.519678 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ca2bfab-d801-4cbf-bb97-c0eff17db807-cert\") pod \"openstack-operator-controller-manager-7cfc658b9-vckkj\" (UID: \"3ca2bfab-d801-4cbf-bb97-c0eff17db807\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.565330 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.726496 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" event={"ID":"6c0af6f9-3f50-4620-94a9-9c9262d73fd9","Type":"ContainerStarted","Data":"0c120c92390cb9889fa50ab3a9db08b42b4b8fbf87609485c9e13c697f7dc4dc"} Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.728760 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" event={"ID":"446adb3e-02c1-4aa1-af8b-a038b43a3a61","Type":"ContainerStarted","Data":"079b9cb3754b71b3254c280c30ced93d4496f5d9aa4c853e0f1b8be11965d5b9"} Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.728790 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" event={"ID":"446adb3e-02c1-4aa1-af8b-a038b43a3a61","Type":"ContainerStarted","Data":"d0e85cf8b0ba3fc0e0c3403670c212c22550a046a59366e1f037c95dd437a5e3"} Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.730801 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" podUID="446adb3e-02c1-4aa1-af8b-a038b43a3a61" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.736297 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" event={"ID":"98df2e22-67cc-485b-856d-ce3e358b0bfd","Type":"ContainerStarted","Data":"ce787f1e873c9f6d7d46bd23d62ae3a7a89f8e8905d10660a82e390c520a39f4"} Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.740522 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" podUID="98df2e22-67cc-485b-856d-ce3e358b0bfd" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.740922 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" event={"ID":"57af99f4-85fe-426a-a608-6c46d72cc16f","Type":"ContainerStarted","Data":"4d4681d670a41fe6c0e0699a662994e0deed188186b1cceb00c06cbf1e07ae40"} Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.740944 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" event={"ID":"57af99f4-85fe-426a-a608-6c46d72cc16f","Type":"ContainerStarted","Data":"4b46ab3563a3e941b9bf0ab00e823f922f5a2ae9dde1555f9b2c6a0ee6f1428d"} Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.743228 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" podUID="57af99f4-85fe-426a-a608-6c46d72cc16f" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.745761 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" event={"ID":"57067c6a-4af7-4b0f-93cf-ac544f00a239","Type":"ContainerStarted","Data":"dc358ee8775f4e89f215ed189755081a524448316e46e3c680e7256d80165efe"} Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.746919 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" podUID="57067c6a-4af7-4b0f-93cf-ac544f00a239" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.760392 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" event={"ID":"1a18b889-577c-4bb0-9d34-b2e560318144","Type":"ContainerStarted","Data":"3044b8404ca05696eb0a47bb017e2a650d1eccddc2b3dc2d086d69238ebb66bd"} Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.774323 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" event={"ID":"5ecfad04-5c99-47dd-b1f5-782b0aea4061","Type":"ContainerStarted","Data":"5e3bf4a5b2b99f09dd9f72c0a324d03c39d38d2440d1c6e47bac2493edbc67c1"} Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.774395 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" event={"ID":"5ecfad04-5c99-47dd-b1f5-782b0aea4061","Type":"ContainerStarted","Data":"44d5ff503efe00409762fc451ce6408628b724692d994c959d1b9161a685e4e6"} Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.785962 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" podUID="5ecfad04-5c99-47dd-b1f5-782b0aea4061" Oct 05 07:09:50 crc kubenswrapper[4935]: E1005 07:09:50.793228 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" podUID="8a31c4d9-3e39-4c59-ac21-f202677474ee" Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.805087 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" event={"ID":"8a31c4d9-3e39-4c59-ac21-f202677474ee","Type":"ContainerStarted","Data":"e4f5351200a284d80a6a61fc333841c2121aded899a8c5c1b29194a3b12879ae"} Oct 05 07:09:50 crc kubenswrapper[4935]: I1005 07:09:50.805134 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" event={"ID":"8a31c4d9-3e39-4c59-ac21-f202677474ee","Type":"ContainerStarted","Data":"8bbac22b0a86e4bf0cb73f07175fdb57cf9fd3cfeac0ca0a807465d23c09cbf9"} Oct 05 07:09:51 crc kubenswrapper[4935]: I1005 07:09:51.124568 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj"] Oct 05 07:09:51 crc kubenswrapper[4935]: W1005 07:09:51.149859 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ca2bfab_d801_4cbf_bb97_c0eff17db807.slice/crio-c2ed2be356cea1445ce51be462809da2f33b776a91d840566aa517630801465c WatchSource:0}: Error finding container c2ed2be356cea1445ce51be462809da2f33b776a91d840566aa517630801465c: Status 404 returned error can't find the container with id c2ed2be356cea1445ce51be462809da2f33b776a91d840566aa517630801465c Oct 05 07:09:51 crc kubenswrapper[4935]: I1005 07:09:51.805602 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" event={"ID":"3ca2bfab-d801-4cbf-bb97-c0eff17db807","Type":"ContainerStarted","Data":"e3da78a34db7011e3d98f8a3d5925e85c2a63a05845febf7bfa5f032655c4cf7"} Oct 05 07:09:51 crc kubenswrapper[4935]: I1005 07:09:51.805841 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" event={"ID":"3ca2bfab-d801-4cbf-bb97-c0eff17db807","Type":"ContainerStarted","Data":"c2ed2be356cea1445ce51be462809da2f33b776a91d840566aa517630801465c"} Oct 05 07:09:51 crc kubenswrapper[4935]: E1005 07:09:51.807435 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" podUID="98df2e22-67cc-485b-856d-ce3e358b0bfd" Oct 05 07:09:51 crc kubenswrapper[4935]: E1005 07:09:51.807626 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" podUID="57af99f4-85fe-426a-a608-6c46d72cc16f" Oct 05 07:09:51 crc kubenswrapper[4935]: E1005 07:09:51.807629 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" podUID="8a31c4d9-3e39-4c59-ac21-f202677474ee" Oct 05 07:09:51 crc kubenswrapper[4935]: E1005 07:09:51.807670 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" podUID="5ecfad04-5c99-47dd-b1f5-782b0aea4061" Oct 05 07:09:51 crc kubenswrapper[4935]: E1005 07:09:51.808124 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" podUID="446adb3e-02c1-4aa1-af8b-a038b43a3a61" Oct 05 07:09:51 crc kubenswrapper[4935]: E1005 07:09:51.808791 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" podUID="57067c6a-4af7-4b0f-93cf-ac544f00a239" Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.884541 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" event={"ID":"5b24bc52-99be-499c-90ac-91114405b8f6","Type":"ContainerStarted","Data":"d819cd6b29401852a7ef6efe4bdd27474e70914147e8fa18125009ec41694b75"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.889082 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" event={"ID":"1fb98234-7f3b-491f-871a-106ce0f2af2a","Type":"ContainerStarted","Data":"8ef66d17baa9747562bf21370db5f0826bc987cd8a6f133643e8a56dfa758a50"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.895074 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" event={"ID":"1ef01eb4-46b6-4859-a53f-1b9b58f9f408","Type":"ContainerStarted","Data":"c069c2b8a166efc20f2fab1bdf4e5afe98200e26814aa77127e5046819824240"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.914172 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" event={"ID":"1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e","Type":"ContainerStarted","Data":"23bfb08c8ff1bb5bf0ba7649c6aa83fffff3442b589b9c073988f03d7e882f6c"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.929167 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" event={"ID":"3ca2bfab-d801-4cbf-bb97-c0eff17db807","Type":"ContainerStarted","Data":"57b0dbe9ca7e7323792cb67a9a92994690650bf0f53f23305827a63391227491"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.930004 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.933062 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" event={"ID":"dd907a5c-8a2c-459a-b4e0-2f1341fe956b","Type":"ContainerStarted","Data":"535f5f41dac4779449915d76cdcf9ff27ac340b603d9384d135546252393ccac"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.937148 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.969506 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-vckkj" podStartSLOduration=11.969491523 podStartE2EDuration="11.969491523s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:59.963966786 +0000 UTC m=+1033.846593246" watchObservedRunningTime="2025-10-05 07:09:59.969491523 +0000 UTC m=+1033.852117983" Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.970645 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" event={"ID":"317d3174-23c3-4af7-b289-26844e8d3f12","Type":"ContainerStarted","Data":"b84e5af709b4820ce16ed34a248065c8368dd2f1eb6dd62e0125324d854685cc"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.971366 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.977705 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" event={"ID":"451246ff-f04c-4b5b-a956-41592f58ddb9","Type":"ContainerStarted","Data":"fd3c9f5ac530aca8596adb3a6128c77122043dc8473f6786cfa9f789ca5c1d12"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.978302 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.979723 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" event={"ID":"96b524ed-4a5d-4c95-b076-c7f68fab13a2","Type":"ContainerStarted","Data":"588ec234ffd62b60cf62599da4a1ed756412e40549f8b1bb08824af78223ba6a"} Oct 05 07:09:59 crc kubenswrapper[4935]: I1005 07:09:59.986740 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" event={"ID":"1a18b889-577c-4bb0-9d34-b2e560318144","Type":"ContainerStarted","Data":"ba8cf7f00aa07b6449f4add434ee4aa52f4a1c3c6d740057f524c7708527341f"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.003667 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" event={"ID":"6c0af6f9-3f50-4620-94a9-9c9262d73fd9","Type":"ContainerStarted","Data":"8c13fd8c45df56d1b8d4e018a3c1a7971bd599c535f2e0ebf017e1653f9c435f"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.021555 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" event={"ID":"c7430e4b-67b4-45b5-8444-1ee6cf758c5b","Type":"ContainerStarted","Data":"ca893b417922b40de9b0a633071302a5c285d5ea8ff594e5d58e629a8b407c43"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.021597 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" event={"ID":"c7430e4b-67b4-45b5-8444-1ee6cf758c5b","Type":"ContainerStarted","Data":"4d169dfd3d5a069d98c9ef1eea901034bf21e2a2b2474577cbb16c63db996f66"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.022212 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.050447 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" podStartSLOduration=3.410796179 podStartE2EDuration="13.05042957s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:48.979961164 +0000 UTC m=+1022.862587624" lastFinishedPulling="2025-10-05 07:09:58.619594555 +0000 UTC m=+1032.502221015" observedRunningTime="2025-10-05 07:10:00.043685361 +0000 UTC m=+1033.926311821" watchObservedRunningTime="2025-10-05 07:10:00.05042957 +0000 UTC m=+1033.933056030" Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.071995 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" event={"ID":"3078f420-c252-4650-ab0a-17c92ab8d1c6","Type":"ContainerStarted","Data":"095a4533a38dbafcdd076645ce89064858f9a7ef59b2946c2c99778f080155a5"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.080425 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" podStartSLOduration=4.01438378 podStartE2EDuration="13.080402655s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.556149348 +0000 UTC m=+1023.438775808" lastFinishedPulling="2025-10-05 07:09:58.622168233 +0000 UTC m=+1032.504794683" observedRunningTime="2025-10-05 07:10:00.067819321 +0000 UTC m=+1033.950445781" watchObservedRunningTime="2025-10-05 07:10:00.080402655 +0000 UTC m=+1033.963029115" Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.118145 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" event={"ID":"39f750d8-d3c5-4164-ab98-12c031e752ab","Type":"ContainerStarted","Data":"0fbd2af2656edb3024f655452078b043f59d2d677d8d70e1de0fdf641a15f92b"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.126067 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" event={"ID":"eb1aa07b-d43b-4ad3-803c-042de10feab8","Type":"ContainerStarted","Data":"f4e835cbcf584922bd05a9592777cd2f7e6efe7e63318e3a7f4551079d3f5ec9"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.126110 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" event={"ID":"eb1aa07b-d43b-4ad3-803c-042de10feab8","Type":"ContainerStarted","Data":"e22c8ca3b7edac3528c3cfb83b62af561c93ef62b1f1a17bd3e46d09e67ee1b5"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.126941 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.164311 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" event={"ID":"8ca55603-341c-4a88-aaec-34ce00cec8fb","Type":"ContainerStarted","Data":"9add5dbc74d7b59f81b3ec457f2256a46c5078f28deba34c881ff81a944d14d0"} Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.166034 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" podStartSLOduration=3.917389736 podStartE2EDuration="13.166015076s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.361411592 +0000 UTC m=+1023.244038052" lastFinishedPulling="2025-10-05 07:09:58.610036932 +0000 UTC m=+1032.492663392" observedRunningTime="2025-10-05 07:10:00.094202941 +0000 UTC m=+1033.976829391" watchObservedRunningTime="2025-10-05 07:10:00.166015076 +0000 UTC m=+1034.048641526" Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.167578 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" podStartSLOduration=3.937496659 podStartE2EDuration="13.167572307s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.378946277 +0000 UTC m=+1023.261572737" lastFinishedPulling="2025-10-05 07:09:58.609021925 +0000 UTC m=+1032.491648385" observedRunningTime="2025-10-05 07:10:00.15335098 +0000 UTC m=+1034.035977440" watchObservedRunningTime="2025-10-05 07:10:00.167572307 +0000 UTC m=+1034.050198767" Oct 05 07:10:00 crc kubenswrapper[4935]: I1005 07:10:00.175981 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" event={"ID":"1201b72a-0267-4e34-bdf5-6ddb0f9299dd","Type":"ContainerStarted","Data":"cde15dfd03a4158e01ed7d9bea328444fa8c1dbbcf75c4ff9ea8621e7487f524"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.182947 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" event={"ID":"8ca55603-341c-4a88-aaec-34ce00cec8fb","Type":"ContainerStarted","Data":"5230ac2ec8ee321de550d204c383af958dadb14e850efb04d7b21f5c42130fec"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.183450 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.184829 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" event={"ID":"1a18b889-577c-4bb0-9d34-b2e560318144","Type":"ContainerStarted","Data":"b0ed11e1b087d05dc5eb46ff1cc6bc3779ca4d985aab7ada82f1998f3eeec585"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.185190 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.186618 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" event={"ID":"dd907a5c-8a2c-459a-b4e0-2f1341fe956b","Type":"ContainerStarted","Data":"1703efb127b388dd3e45836d12bd4bee58bafe9734981c17e2d4a7a1d6685f2d"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.186756 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.188296 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" event={"ID":"451246ff-f04c-4b5b-a956-41592f58ddb9","Type":"ContainerStarted","Data":"1ed4f17889e094001f97a8bfbf9ad385e4abf9e5b05ab26dafa6391b1f461f8f"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.190008 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" event={"ID":"96b524ed-4a5d-4c95-b076-c7f68fab13a2","Type":"ContainerStarted","Data":"971eecac712d28c94c89191f76678f395ddde05a59b298c32ba595e4b2a737c0"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.190145 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.191546 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" event={"ID":"1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e","Type":"ContainerStarted","Data":"cd2314b9a81d132e92b9f6f1f5d0ffc8cfd057c47e0940244077b4e23a28bce7"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.191697 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.192853 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" event={"ID":"5b24bc52-99be-499c-90ac-91114405b8f6","Type":"ContainerStarted","Data":"d7bbb46d45534ea9aed38806ab7cd60c88ddfc10cc2a7e949200e81575d21272"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.192933 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.194599 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" event={"ID":"317d3174-23c3-4af7-b289-26844e8d3f12","Type":"ContainerStarted","Data":"4bb064bc7710062300ce7bd09b7bfe7b250c2cae662889e7efe27b27cd39f1d8"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.196139 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" event={"ID":"1ef01eb4-46b6-4859-a53f-1b9b58f9f408","Type":"ContainerStarted","Data":"0ca312f486ce25e69b00df3832af9c8cd5ea639a776005331b7c381035042445"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.196258 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.197666 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" event={"ID":"3078f420-c252-4650-ab0a-17c92ab8d1c6","Type":"ContainerStarted","Data":"f2c5720822cc5f16f6814dfe7ab5f1aa216b2aa52d63e9acd68b533d420bf316"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.197756 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.199233 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" event={"ID":"39f750d8-d3c5-4164-ab98-12c031e752ab","Type":"ContainerStarted","Data":"f05dfa7346a9b270aa08816e6f27482214889b132a35640b892b916f7ac15d9c"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.199384 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.200515 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" event={"ID":"1201b72a-0267-4e34-bdf5-6ddb0f9299dd","Type":"ContainerStarted","Data":"ef6d94af0c576ae1254c125b5787723d88ebf4c88571443a041b5f1b1ef8b76b"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.200559 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.201797 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" event={"ID":"6c0af6f9-3f50-4620-94a9-9c9262d73fd9","Type":"ContainerStarted","Data":"d288da7a1bba97862592b86805db217c1cd5f392fbf8e28b18af0ecce90b555a"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.201918 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.203050 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" event={"ID":"1fb98234-7f3b-491f-871a-106ce0f2af2a","Type":"ContainerStarted","Data":"ff3d49826b6fa831610adace500774a6593b74614e65fc86fc177a45b2605f1f"} Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.203199 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.209403 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" podStartSLOduration=4.212627013 podStartE2EDuration="13.209383871s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.622871608 +0000 UTC m=+1023.505498068" lastFinishedPulling="2025-10-05 07:09:58.619628466 +0000 UTC m=+1032.502254926" observedRunningTime="2025-10-05 07:10:01.204240845 +0000 UTC m=+1035.086867305" watchObservedRunningTime="2025-10-05 07:10:01.209383871 +0000 UTC m=+1035.092010331" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.245765 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" podStartSLOduration=4.995200951 podStartE2EDuration="13.245747166s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="2025-10-05 07:09:50.387466209 +0000 UTC m=+1024.270092679" lastFinishedPulling="2025-10-05 07:09:58.638012434 +0000 UTC m=+1032.520638894" observedRunningTime="2025-10-05 07:10:01.234740184 +0000 UTC m=+1035.117366634" watchObservedRunningTime="2025-10-05 07:10:01.245747166 +0000 UTC m=+1035.128373626" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.255288 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" podStartSLOduration=5.43490247 podStartE2EDuration="14.255269399s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.788647625 +0000 UTC m=+1023.671274085" lastFinishedPulling="2025-10-05 07:09:58.609014534 +0000 UTC m=+1032.491641014" observedRunningTime="2025-10-05 07:10:01.253135082 +0000 UTC m=+1035.135761542" watchObservedRunningTime="2025-10-05 07:10:01.255269399 +0000 UTC m=+1035.137895859" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.271781 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" podStartSLOduration=5.220188333 podStartE2EDuration="14.271764366s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.558396547 +0000 UTC m=+1023.441023007" lastFinishedPulling="2025-10-05 07:09:58.60997254 +0000 UTC m=+1032.492599040" observedRunningTime="2025-10-05 07:10:01.269124936 +0000 UTC m=+1035.151751396" watchObservedRunningTime="2025-10-05 07:10:01.271764366 +0000 UTC m=+1035.154390826" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.288982 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" podStartSLOduration=4.811099703 podStartE2EDuration="14.28886701s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.132256564 +0000 UTC m=+1023.014883034" lastFinishedPulling="2025-10-05 07:09:58.610023881 +0000 UTC m=+1032.492650341" observedRunningTime="2025-10-05 07:10:01.285040558 +0000 UTC m=+1035.167667028" watchObservedRunningTime="2025-10-05 07:10:01.28886701 +0000 UTC m=+1035.171493470" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.305512 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" podStartSLOduration=5.306125053 podStartE2EDuration="14.305495411s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.621757058 +0000 UTC m=+1023.504383518" lastFinishedPulling="2025-10-05 07:09:58.621127416 +0000 UTC m=+1032.503753876" observedRunningTime="2025-10-05 07:10:01.304957367 +0000 UTC m=+1035.187583837" watchObservedRunningTime="2025-10-05 07:10:01.305495411 +0000 UTC m=+1035.188121871" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.319152 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" podStartSLOduration=5.333990753 podStartE2EDuration="14.319136993s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.634586519 +0000 UTC m=+1023.517212979" lastFinishedPulling="2025-10-05 07:09:58.619732759 +0000 UTC m=+1032.502359219" observedRunningTime="2025-10-05 07:10:01.317084208 +0000 UTC m=+1035.199710668" watchObservedRunningTime="2025-10-05 07:10:01.319136993 +0000 UTC m=+1035.201763453" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.338516 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" podStartSLOduration=5.337606098 podStartE2EDuration="14.338495726s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.619116468 +0000 UTC m=+1023.501742928" lastFinishedPulling="2025-10-05 07:09:58.620006096 +0000 UTC m=+1032.502632556" observedRunningTime="2025-10-05 07:10:01.335655861 +0000 UTC m=+1035.218282321" watchObservedRunningTime="2025-10-05 07:10:01.338495726 +0000 UTC m=+1035.221122186" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.353041 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" podStartSLOduration=4.508035414 podStartE2EDuration="14.353025212s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:48.76416601 +0000 UTC m=+1022.646792470" lastFinishedPulling="2025-10-05 07:09:58.609155808 +0000 UTC m=+1032.491782268" observedRunningTime="2025-10-05 07:10:01.349445627 +0000 UTC m=+1035.232072087" watchObservedRunningTime="2025-10-05 07:10:01.353025212 +0000 UTC m=+1035.235651672" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.366730 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" podStartSLOduration=4.681184945 podStartE2EDuration="14.366696914s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:48.923787163 +0000 UTC m=+1022.806413623" lastFinishedPulling="2025-10-05 07:09:58.609299132 +0000 UTC m=+1032.491925592" observedRunningTime="2025-10-05 07:10:01.366428717 +0000 UTC m=+1035.249055177" watchObservedRunningTime="2025-10-05 07:10:01.366696914 +0000 UTC m=+1035.249323414" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.388339 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" podStartSLOduration=5.334494307 podStartE2EDuration="14.388306198s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.556666712 +0000 UTC m=+1023.439293172" lastFinishedPulling="2025-10-05 07:09:58.610478603 +0000 UTC m=+1032.493105063" observedRunningTime="2025-10-05 07:10:01.38237065 +0000 UTC m=+1035.264997110" watchObservedRunningTime="2025-10-05 07:10:01.388306198 +0000 UTC m=+1035.270932658" Oct 05 07:10:01 crc kubenswrapper[4935]: I1005 07:10:01.401698 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" podStartSLOduration=5.124421204 podStartE2EDuration="14.401671652s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.331727025 +0000 UTC m=+1023.214353485" lastFinishedPulling="2025-10-05 07:09:58.608977473 +0000 UTC m=+1032.491603933" observedRunningTime="2025-10-05 07:10:01.401343543 +0000 UTC m=+1035.283970033" watchObservedRunningTime="2025-10-05 07:10:01.401671652 +0000 UTC m=+1035.284298142" Oct 05 07:10:05 crc kubenswrapper[4935]: I1005 07:10:05.244936 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" event={"ID":"57af99f4-85fe-426a-a608-6c46d72cc16f","Type":"ContainerStarted","Data":"f85454f5eade75d848bc9f69f4f02b2ae7ff857553ee36cfa18d9fe82dd67ce1"} Oct 05 07:10:05 crc kubenswrapper[4935]: I1005 07:10:05.245805 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" Oct 05 07:10:05 crc kubenswrapper[4935]: I1005 07:10:05.246362 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" event={"ID":"5ecfad04-5c99-47dd-b1f5-782b0aea4061","Type":"ContainerStarted","Data":"ee6f55ec0a2c2b0458693b436f9cfac2553ba6a7301d07551266e70ff538405d"} Oct 05 07:10:05 crc kubenswrapper[4935]: I1005 07:10:05.246564 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" Oct 05 07:10:05 crc kubenswrapper[4935]: I1005 07:10:05.262149 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" podStartSLOduration=2.598380324 podStartE2EDuration="17.262121804s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.829684294 +0000 UTC m=+1023.712310754" lastFinishedPulling="2025-10-05 07:10:04.493425774 +0000 UTC m=+1038.376052234" observedRunningTime="2025-10-05 07:10:05.259464574 +0000 UTC m=+1039.142091034" watchObservedRunningTime="2025-10-05 07:10:05.262121804 +0000 UTC m=+1039.144748254" Oct 05 07:10:07 crc kubenswrapper[4935]: I1005 07:10:07.276797 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" event={"ID":"446adb3e-02c1-4aa1-af8b-a038b43a3a61","Type":"ContainerStarted","Data":"8ee0781851af238cdf2f500789d3ee362030c16a324d1307d2a429aff372ae7e"} Oct 05 07:10:07 crc kubenswrapper[4935]: I1005 07:10:07.277040 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" Oct 05 07:10:07 crc kubenswrapper[4935]: I1005 07:10:07.280740 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" event={"ID":"8a31c4d9-3e39-4c59-ac21-f202677474ee","Type":"ContainerStarted","Data":"b5b80bee2db333cd20d537377b912284a1aa97fd7a55dfc8946b4fd9e375c246"} Oct 05 07:10:07 crc kubenswrapper[4935]: I1005 07:10:07.280944 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" Oct 05 07:10:07 crc kubenswrapper[4935]: I1005 07:10:07.295253 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" podStartSLOduration=4.722248462 podStartE2EDuration="19.295234665s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.912523641 +0000 UTC m=+1023.795150101" lastFinishedPulling="2025-10-05 07:10:04.485509844 +0000 UTC m=+1038.368136304" observedRunningTime="2025-10-05 07:10:05.279378052 +0000 UTC m=+1039.162004532" watchObservedRunningTime="2025-10-05 07:10:07.295234665 +0000 UTC m=+1041.177861125" Oct 05 07:10:07 crc kubenswrapper[4935]: I1005 07:10:07.295931 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" podStartSLOduration=2.9565169449999997 podStartE2EDuration="19.295925314s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.795034515 +0000 UTC m=+1023.677660975" lastFinishedPulling="2025-10-05 07:10:06.134442884 +0000 UTC m=+1040.017069344" observedRunningTime="2025-10-05 07:10:07.291769153 +0000 UTC m=+1041.174395623" watchObservedRunningTime="2025-10-05 07:10:07.295925314 +0000 UTC m=+1041.178551774" Oct 05 07:10:07 crc kubenswrapper[4935]: I1005 07:10:07.312686 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" podStartSLOduration=3.009734347 podStartE2EDuration="19.312671298s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.81632132 +0000 UTC m=+1023.698947780" lastFinishedPulling="2025-10-05 07:10:06.119258271 +0000 UTC m=+1040.001884731" observedRunningTime="2025-10-05 07:10:07.310631304 +0000 UTC m=+1041.193257764" watchObservedRunningTime="2025-10-05 07:10:07.312671298 +0000 UTC m=+1041.195297758" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.063659 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-6nxjj" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.071591 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-nrs2t" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.100283 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-ntjg7" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.153733 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-55r29" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.206224 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-s8gzx" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.216999 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-l9bf8" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.231343 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zk5fs" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.247870 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-d2l5q" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.274728 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-g84c7" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.303049 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" event={"ID":"98df2e22-67cc-485b-856d-ce3e358b0bfd","Type":"ContainerStarted","Data":"0ba611106ac5213f27bd22f6e2b8ee7c9c81e070574487b09177406e22744ee5"} Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.319014 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt" podStartSLOduration=2.726713378 podStartE2EDuration="20.318995771s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.921472789 +0000 UTC m=+1023.804099249" lastFinishedPulling="2025-10-05 07:10:07.513755182 +0000 UTC m=+1041.396381642" observedRunningTime="2025-10-05 07:10:08.3170894 +0000 UTC m=+1042.199715870" watchObservedRunningTime="2025-10-05 07:10:08.318995771 +0000 UTC m=+1042.201622231" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.331514 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-22j5g" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.369719 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-lhzb8" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.390261 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-llst6" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.489954 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-55k65" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.764853 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8xv6f" Oct 05 07:10:08 crc kubenswrapper[4935]: I1005 07:10:08.850192 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-b285k" Oct 05 07:10:09 crc kubenswrapper[4935]: I1005 07:10:09.310759 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" event={"ID":"57067c6a-4af7-4b0f-93cf-ac544f00a239","Type":"ContainerStarted","Data":"6c16dbb48168942f35d24ddd76cceb0f8c19a3bd027712c209ad4f3c1b23c79c"} Oct 05 07:10:09 crc kubenswrapper[4935]: I1005 07:10:09.311092 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" Oct 05 07:10:09 crc kubenswrapper[4935]: I1005 07:10:09.335291 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" podStartSLOduration=2.293635041 podStartE2EDuration="21.335272569s" podCreationTimestamp="2025-10-05 07:09:48 +0000 UTC" firstStartedPulling="2025-10-05 07:09:49.63879102 +0000 UTC m=+1023.521417470" lastFinishedPulling="2025-10-05 07:10:08.680428538 +0000 UTC m=+1042.563054998" observedRunningTime="2025-10-05 07:10:09.32701062 +0000 UTC m=+1043.209637130" watchObservedRunningTime="2025-10-05 07:10:09.335272569 +0000 UTC m=+1043.217899049" Oct 05 07:10:09 crc kubenswrapper[4935]: I1005 07:10:09.830905 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz" Oct 05 07:10:14 crc kubenswrapper[4935]: I1005 07:10:14.289509 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:10:14 crc kubenswrapper[4935]: I1005 07:10:14.289599 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:10:18 crc kubenswrapper[4935]: I1005 07:10:18.573948 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-g2828" Oct 05 07:10:18 crc kubenswrapper[4935]: I1005 07:10:18.833770 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-xkph9" Oct 05 07:10:18 crc kubenswrapper[4935]: I1005 07:10:18.897281 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-hg5ns" Oct 05 07:10:18 crc kubenswrapper[4935]: I1005 07:10:18.945114 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zcr8x" Oct 05 07:10:18 crc kubenswrapper[4935]: I1005 07:10:18.988147 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-rtvd8" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.418342 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-qfggd"] Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.420350 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.422909 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.423161 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.423274 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.423380 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-qjqnt" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.425452 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-qfggd"] Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.440735 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98zzb\" (UniqueName: \"kubernetes.io/projected/e6ac4b62-a353-484c-ac04-77c77963cf9f-kube-api-access-98zzb\") pod \"dnsmasq-dns-546d69f86c-qfggd\" (UID: \"e6ac4b62-a353-484c-ac04-77c77963cf9f\") " pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.440852 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6ac4b62-a353-484c-ac04-77c77963cf9f-config\") pod \"dnsmasq-dns-546d69f86c-qfggd\" (UID: \"e6ac4b62-a353-484c-ac04-77c77963cf9f\") " pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.490400 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-6rkzk"] Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.492232 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.494098 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.501996 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-6rkzk"] Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.541797 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-config\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.541907 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6ac4b62-a353-484c-ac04-77c77963cf9f-config\") pod \"dnsmasq-dns-546d69f86c-qfggd\" (UID: \"e6ac4b62-a353-484c-ac04-77c77963cf9f\") " pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.541971 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.542010 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzpg8\" (UniqueName: \"kubernetes.io/projected/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-kube-api-access-xzpg8\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.542069 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98zzb\" (UniqueName: \"kubernetes.io/projected/e6ac4b62-a353-484c-ac04-77c77963cf9f-kube-api-access-98zzb\") pod \"dnsmasq-dns-546d69f86c-qfggd\" (UID: \"e6ac4b62-a353-484c-ac04-77c77963cf9f\") " pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.543543 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6ac4b62-a353-484c-ac04-77c77963cf9f-config\") pod \"dnsmasq-dns-546d69f86c-qfggd\" (UID: \"e6ac4b62-a353-484c-ac04-77c77963cf9f\") " pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.560850 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98zzb\" (UniqueName: \"kubernetes.io/projected/e6ac4b62-a353-484c-ac04-77c77963cf9f-kube-api-access-98zzb\") pod \"dnsmasq-dns-546d69f86c-qfggd\" (UID: \"e6ac4b62-a353-484c-ac04-77c77963cf9f\") " pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.643000 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-config\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.643111 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.643147 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzpg8\" (UniqueName: \"kubernetes.io/projected/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-kube-api-access-xzpg8\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.644452 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-config\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.645103 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.663258 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzpg8\" (UniqueName: \"kubernetes.io/projected/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-kube-api-access-xzpg8\") pod \"dnsmasq-dns-7f9579fb85-6rkzk\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.744859 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:33 crc kubenswrapper[4935]: I1005 07:10:33.821501 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:34 crc kubenswrapper[4935]: I1005 07:10:34.165065 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-qfggd"] Oct 05 07:10:34 crc kubenswrapper[4935]: W1005 07:10:34.165687 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6ac4b62_a353_484c_ac04_77c77963cf9f.slice/crio-75d8f3431302dc46f4771338bb89e2e201aacfa75604d8aa150a6fc647b11817 WatchSource:0}: Error finding container 75d8f3431302dc46f4771338bb89e2e201aacfa75604d8aa150a6fc647b11817: Status 404 returned error can't find the container with id 75d8f3431302dc46f4771338bb89e2e201aacfa75604d8aa150a6fc647b11817 Oct 05 07:10:34 crc kubenswrapper[4935]: I1005 07:10:34.247611 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-6rkzk"] Oct 05 07:10:34 crc kubenswrapper[4935]: W1005 07:10:34.249779 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb72eb23_93d7_4f24_aecf_e4f83ffaf24c.slice/crio-0ec6dd9e56d889d7697817b6bce31af365fc0b86c5726bf870445edc84c7aa84 WatchSource:0}: Error finding container 0ec6dd9e56d889d7697817b6bce31af365fc0b86c5726bf870445edc84c7aa84: Status 404 returned error can't find the container with id 0ec6dd9e56d889d7697817b6bce31af365fc0b86c5726bf870445edc84c7aa84 Oct 05 07:10:34 crc kubenswrapper[4935]: I1005 07:10:34.487370 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" event={"ID":"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c","Type":"ContainerStarted","Data":"0ec6dd9e56d889d7697817b6bce31af365fc0b86c5726bf870445edc84c7aa84"} Oct 05 07:10:34 crc kubenswrapper[4935]: I1005 07:10:34.488848 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-546d69f86c-qfggd" event={"ID":"e6ac4b62-a353-484c-ac04-77c77963cf9f","Type":"ContainerStarted","Data":"75d8f3431302dc46f4771338bb89e2e201aacfa75604d8aa150a6fc647b11817"} Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.571418 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-qfggd"] Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.591995 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bf999f689-rksqm"] Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.598967 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.602778 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bf999f689-rksqm"] Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.670741 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-config\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.670793 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6qzb\" (UniqueName: \"kubernetes.io/projected/1553404c-fd63-4399-86bc-7b8550b8c372-kube-api-access-f6qzb\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.670848 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-dns-svc\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.771536 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-config\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.771585 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6qzb\" (UniqueName: \"kubernetes.io/projected/1553404c-fd63-4399-86bc-7b8550b8c372-kube-api-access-f6qzb\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.771660 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-dns-svc\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.772497 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-dns-svc\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.772525 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-config\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.793172 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6qzb\" (UniqueName: \"kubernetes.io/projected/1553404c-fd63-4399-86bc-7b8550b8c372-kube-api-access-f6qzb\") pod \"dnsmasq-dns-bf999f689-rksqm\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:35 crc kubenswrapper[4935]: I1005 07:10:35.927173 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.171195 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-6rkzk"] Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.196443 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-pgq4m"] Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.198087 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.210156 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-pgq4m"] Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.280239 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgqk9\" (UniqueName: \"kubernetes.io/projected/18a20547-c4f6-418d-ad4d-61116540d869-kube-api-access-pgqk9\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.280312 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-config\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.280444 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.391609 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.392421 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.392554 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgqk9\" (UniqueName: \"kubernetes.io/projected/18a20547-c4f6-418d-ad4d-61116540d869-kube-api-access-pgqk9\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.392863 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-config\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.393451 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-config\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.410126 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgqk9\" (UniqueName: \"kubernetes.io/projected/18a20547-c4f6-418d-ad4d-61116540d869-kube-api-access-pgqk9\") pod \"dnsmasq-dns-5d4d9f7875-pgq4m\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.483751 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bf999f689-rksqm"] Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.518829 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.737420 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.739328 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.741651 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.742019 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ltjw5" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.742047 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.742401 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.742667 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.742723 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.742982 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.745769 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903203 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903246 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903266 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9vhq\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-kube-api-access-v9vhq\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903287 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903491 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903591 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53fe24e1-5efb-4422-ade0-16e315b7e328-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903620 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903643 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903669 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903695 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53fe24e1-5efb-4422-ade0-16e315b7e328-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.903738 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:36 crc kubenswrapper[4935]: I1005 07:10:36.991512 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-pgq4m"] Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008601 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53fe24e1-5efb-4422-ade0-16e315b7e328-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008641 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008759 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008785 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008804 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53fe24e1-5efb-4422-ade0-16e315b7e328-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008828 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008880 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008919 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008938 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9vhq\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-kube-api-access-v9vhq\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008958 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.008988 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.009604 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.009960 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.010500 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.010720 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.013135 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.015288 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53fe24e1-5efb-4422-ade0-16e315b7e328-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.015640 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.019036 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53fe24e1-5efb-4422-ade0-16e315b7e328-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.020042 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.020721 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.035351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.035616 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9vhq\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-kube-api-access-v9vhq\") pod \"rabbitmq-cell1-server-0\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.065249 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.329751 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.331825 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.335880 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.336122 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.336574 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.338707 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.338733 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.338960 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-c25cf" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.338966 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.340729 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.519200 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.519960 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8dcf22e2-2acd-4b49-a2bc-813a50f89409-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520013 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520036 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520197 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520245 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520280 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bhxr\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-kube-api-access-2bhxr\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520426 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8dcf22e2-2acd-4b49-a2bc-813a50f89409-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520466 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520494 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.520527 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.538124 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf999f689-rksqm" event={"ID":"1553404c-fd63-4399-86bc-7b8550b8c372","Type":"ContainerStarted","Data":"0a926d1b8e37c004def352a3ab316a544f5c30a6a52e5abd86dda8b89b04f61a"} Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.540648 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" event={"ID":"18a20547-c4f6-418d-ad4d-61116540d869","Type":"ContainerStarted","Data":"2e91a7225d275977c5c162cdb7ce87691a935b3b8c57274eac9bc6617b74a17d"} Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.542310 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622478 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622609 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bhxr\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-kube-api-access-2bhxr\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622673 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8dcf22e2-2acd-4b49-a2bc-813a50f89409-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622691 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622717 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622734 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622764 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622817 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8dcf22e2-2acd-4b49-a2bc-813a50f89409-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622843 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622883 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.622945 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.625278 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.625734 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.625744 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.626274 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.626338 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.627270 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.640353 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8dcf22e2-2acd-4b49-a2bc-813a50f89409-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.640360 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.640408 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.640751 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8dcf22e2-2acd-4b49-a2bc-813a50f89409-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.644746 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bhxr\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-kube-api-access-2bhxr\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.653141 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " pod="openstack/rabbitmq-server-0" Oct 05 07:10:37 crc kubenswrapper[4935]: I1005 07:10:37.957310 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.746865 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.748355 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.750437 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.756350 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.756504 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.756677 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-kd6h7" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.756747 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.758538 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.764520 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843155 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843211 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kolla-config\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843299 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmm7l\" (UniqueName: \"kubernetes.io/projected/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kube-api-access-lmm7l\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843488 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843585 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-default\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843666 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-secrets\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843685 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843802 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.843830 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.944735 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.944797 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kolla-config\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.944849 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmm7l\" (UniqueName: \"kubernetes.io/projected/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kube-api-access-lmm7l\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.944884 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.944918 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-default\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.944957 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-secrets\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.944975 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.944992 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.945013 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.945217 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.946211 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kolla-config\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.946223 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-default\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.946506 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.946597 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.965060 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.965325 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.965331 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-secrets\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.972537 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmm7l\" (UniqueName: \"kubernetes.io/projected/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kube-api-access-lmm7l\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:38 crc kubenswrapper[4935]: I1005 07:10:38.972767 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " pod="openstack/openstack-galera-0" Oct 05 07:10:39 crc kubenswrapper[4935]: I1005 07:10:39.073811 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.119748 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.122277 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.126727 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.126751 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.126881 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xkt8k" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.127457 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.133178 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.268971 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.269071 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.269134 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.269174 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5mws\" (UniqueName: \"kubernetes.io/projected/7bedc55a-2396-4249-ac54-a753a67dca72-kube-api-access-z5mws\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.269201 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.269249 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.269280 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.269325 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.269352 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370526 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370590 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370626 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5mws\" (UniqueName: \"kubernetes.io/projected/7bedc55a-2396-4249-ac54-a753a67dca72-kube-api-access-z5mws\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370642 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370673 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370697 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370726 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370746 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.370929 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.371208 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.371357 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.371544 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.372181 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.390660 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.392985 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.394030 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.395659 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-mxgpg" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.395854 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.396885 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.404400 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.404435 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5mws\" (UniqueName: \"kubernetes.io/projected/7bedc55a-2396-4249-ac54-a753a67dca72-kube-api-access-z5mws\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.413213 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.428624 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.437314 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.457602 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.472003 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.472058 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.472170 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-kolla-config\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.472346 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wn64\" (UniqueName: \"kubernetes.io/projected/b82213d6-93fa-48fc-808b-36f902cce15c-kube-api-access-6wn64\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.472375 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-config-data\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.574619 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wn64\" (UniqueName: \"kubernetes.io/projected/b82213d6-93fa-48fc-808b-36f902cce15c-kube-api-access-6wn64\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.574671 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-config-data\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.574726 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.574753 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.574777 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-kolla-config\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.575417 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-kolla-config\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.575789 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-config-data\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.579754 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.596697 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wn64\" (UniqueName: \"kubernetes.io/projected/b82213d6-93fa-48fc-808b-36f902cce15c-kube-api-access-6wn64\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.600512 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " pod="openstack/memcached-0" Oct 05 07:10:40 crc kubenswrapper[4935]: I1005 07:10:40.790083 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 07:10:41 crc kubenswrapper[4935]: I1005 07:10:41.575723 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"53fe24e1-5efb-4422-ade0-16e315b7e328","Type":"ContainerStarted","Data":"13bd2a8878775261a6323e48f9ceab16063a52e65164f2517b9a6afd89f9fdff"} Oct 05 07:10:42 crc kubenswrapper[4935]: I1005 07:10:42.439177 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:10:42 crc kubenswrapper[4935]: I1005 07:10:42.440276 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:10:42 crc kubenswrapper[4935]: I1005 07:10:42.441927 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-rn68c" Oct 05 07:10:42 crc kubenswrapper[4935]: I1005 07:10:42.451833 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:10:42 crc kubenswrapper[4935]: I1005 07:10:42.607100 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b77dh\" (UniqueName: \"kubernetes.io/projected/c0943221-d85f-4adb-a676-ffb7e7406eae-kube-api-access-b77dh\") pod \"kube-state-metrics-0\" (UID: \"c0943221-d85f-4adb-a676-ffb7e7406eae\") " pod="openstack/kube-state-metrics-0" Oct 05 07:10:42 crc kubenswrapper[4935]: I1005 07:10:42.708488 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b77dh\" (UniqueName: \"kubernetes.io/projected/c0943221-d85f-4adb-a676-ffb7e7406eae-kube-api-access-b77dh\") pod \"kube-state-metrics-0\" (UID: \"c0943221-d85f-4adb-a676-ffb7e7406eae\") " pod="openstack/kube-state-metrics-0" Oct 05 07:10:42 crc kubenswrapper[4935]: I1005 07:10:42.745648 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b77dh\" (UniqueName: \"kubernetes.io/projected/c0943221-d85f-4adb-a676-ffb7e7406eae-kube-api-access-b77dh\") pod \"kube-state-metrics-0\" (UID: \"c0943221-d85f-4adb-a676-ffb7e7406eae\") " pod="openstack/kube-state-metrics-0" Oct 05 07:10:42 crc kubenswrapper[4935]: I1005 07:10:42.755219 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:10:44 crc kubenswrapper[4935]: I1005 07:10:44.289824 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:10:44 crc kubenswrapper[4935]: I1005 07:10:44.289868 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.364639 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.366422 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.371881 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.372004 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-hxt2k" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.372651 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.373691 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.373779 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.381982 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.487410 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.487464 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.487507 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.487535 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.487669 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.487846 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.487940 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.488024 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgpdp\" (UniqueName: \"kubernetes.io/projected/d0d984f8-20c4-4b30-b767-c6106aa279cf-kube-api-access-sgpdp\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590158 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590216 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590255 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590280 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590318 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590387 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590425 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590471 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgpdp\" (UniqueName: \"kubernetes.io/projected/d0d984f8-20c4-4b30-b767-c6106aa279cf-kube-api-access-sgpdp\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.590736 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.591045 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.591418 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.591692 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.598291 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.604595 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.608072 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.608980 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgpdp\" (UniqueName: \"kubernetes.io/projected/d0d984f8-20c4-4b30-b767-c6106aa279cf-kube-api-access-sgpdp\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.613690 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.691264 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.897231 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-px6js"] Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.898459 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-px6js" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.900596 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-pmfv7" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.901072 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.901081 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.907961 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-px6js"] Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.922358 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-pl4nq"] Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.926742 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.949737 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pl4nq"] Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.996791 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xds\" (UniqueName: \"kubernetes.io/projected/971ba03f-bdf3-4788-a28d-76ae1c5c658c-kube-api-access-b7xds\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.996836 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.996859 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run-ovn\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.996879 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/971ba03f-bdf3-4788-a28d-76ae1c5c658c-scripts\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.996910 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-log-ovn\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.996944 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-combined-ca-bundle\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:46 crc kubenswrapper[4935]: I1005 07:10:46.996990 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-ovn-controller-tls-certs\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.098681 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.099206 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/971ba03f-bdf3-4788-a28d-76ae1c5c658c-scripts\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.099303 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-log-ovn\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.099386 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-log\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.099470 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-run\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.099554 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcrnr\" (UniqueName: \"kubernetes.io/projected/aac8194b-9784-4eed-8c35-fb65a584b525-kube-api-access-xcrnr\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.099942 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xds\" (UniqueName: \"kubernetes.io/projected/971ba03f-bdf3-4788-a28d-76ae1c5c658c-kube-api-access-b7xds\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.100033 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run-ovn\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.100116 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-lib\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.100290 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-combined-ca-bundle\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.100756 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-ovn-controller-tls-certs\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.100855 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-etc-ovs\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.100966 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac8194b-9784-4eed-8c35-fb65a584b525-scripts\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.099808 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-log-ovn\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.099627 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.100708 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run-ovn\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.104418 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/971ba03f-bdf3-4788-a28d-76ae1c5c658c-scripts\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.105711 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-combined-ca-bundle\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.107492 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-ovn-controller-tls-certs\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.124399 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xds\" (UniqueName: \"kubernetes.io/projected/971ba03f-bdf3-4788-a28d-76ae1c5c658c-kube-api-access-b7xds\") pod \"ovn-controller-px6js\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.202646 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-etc-ovs\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.202999 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac8194b-9784-4eed-8c35-fb65a584b525-scripts\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.203141 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-log\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.203241 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-run\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.203376 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcrnr\" (UniqueName: \"kubernetes.io/projected/aac8194b-9784-4eed-8c35-fb65a584b525-kube-api-access-xcrnr\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.203431 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-run\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.203033 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-etc-ovs\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.203383 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-log\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.205250 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-lib\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.205591 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-lib\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.206578 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac8194b-9784-4eed-8c35-fb65a584b525-scripts\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.222487 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcrnr\" (UniqueName: \"kubernetes.io/projected/aac8194b-9784-4eed-8c35-fb65a584b525-kube-api-access-xcrnr\") pod \"ovn-controller-ovs-pl4nq\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.225769 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-px6js" Oct 05 07:10:47 crc kubenswrapper[4935]: I1005 07:10:47.251848 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:10:48 crc kubenswrapper[4935]: I1005 07:10:48.388014 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:10:49 crc kubenswrapper[4935]: W1005 07:10:49.030860 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dcf22e2_2acd_4b49_a2bc_813a50f89409.slice/crio-923ad154d30fd0bd4a9a24e52741d6b23c1def8493cafe588f96b8b50acf6fda WatchSource:0}: Error finding container 923ad154d30fd0bd4a9a24e52741d6b23c1def8493cafe588f96b8b50acf6fda: Status 404 returned error can't find the container with id 923ad154d30fd0bd4a9a24e52741d6b23c1def8493cafe588f96b8b50acf6fda Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.034599 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:10:49 crc kubenswrapper[4935]: E1005 07:10:49.061694 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 05 07:10:49 crc kubenswrapper[4935]: E1005 07:10:49.061863 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-98zzb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-546d69f86c-qfggd_openstack(e6ac4b62-a353-484c-ac04-77c77963cf9f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:10:49 crc kubenswrapper[4935]: E1005 07:10:49.063450 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-546d69f86c-qfggd" podUID="e6ac4b62-a353-484c-ac04-77c77963cf9f" Oct 05 07:10:49 crc kubenswrapper[4935]: E1005 07:10:49.072812 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 05 07:10:49 crc kubenswrapper[4935]: E1005 07:10:49.073132 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xzpg8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7f9579fb85-6rkzk_openstack(eb72eb23-93d7-4f24-aecf-e4f83ffaf24c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:10:49 crc kubenswrapper[4935]: E1005 07:10:49.075071 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" podUID="eb72eb23-93d7-4f24-aecf-e4f83ffaf24c" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.438365 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.445417 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.445559 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.456012 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.456813 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.461550 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-bldkp" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.468813 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.556690 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.556749 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.556776 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldxzj\" (UniqueName: \"kubernetes.io/projected/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-kube-api-access-ldxzj\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.556925 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.557033 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-config\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.557062 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.557175 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.557297 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.646865 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dcf22e2-2acd-4b49-a2bc-813a50f89409","Type":"ContainerStarted","Data":"923ad154d30fd0bd4a9a24e52741d6b23c1def8493cafe588f96b8b50acf6fda"} Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.664183 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.664394 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.664505 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.664541 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.664558 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldxzj\" (UniqueName: \"kubernetes.io/projected/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-kube-api-access-ldxzj\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.664580 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.664634 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-config\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.664657 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.665315 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.666550 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.666851 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-config\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.667124 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.673023 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.675313 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.676630 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.703850 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.705234 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldxzj\" (UniqueName: \"kubernetes.io/projected/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-kube-api-access-ldxzj\") pod \"ovsdbserver-nb-0\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:49 crc kubenswrapper[4935]: I1005 07:10:49.789356 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.477975 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.527204 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.584213 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98zzb\" (UniqueName: \"kubernetes.io/projected/e6ac4b62-a353-484c-ac04-77c77963cf9f-kube-api-access-98zzb\") pod \"e6ac4b62-a353-484c-ac04-77c77963cf9f\" (UID: \"e6ac4b62-a353-484c-ac04-77c77963cf9f\") " Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.584609 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6ac4b62-a353-484c-ac04-77c77963cf9f-config\") pod \"e6ac4b62-a353-484c-ac04-77c77963cf9f\" (UID: \"e6ac4b62-a353-484c-ac04-77c77963cf9f\") " Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.585857 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6ac4b62-a353-484c-ac04-77c77963cf9f-config" (OuterVolumeSpecName: "config") pod "e6ac4b62-a353-484c-ac04-77c77963cf9f" (UID: "e6ac4b62-a353-484c-ac04-77c77963cf9f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.591338 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6ac4b62-a353-484c-ac04-77c77963cf9f-kube-api-access-98zzb" (OuterVolumeSpecName: "kube-api-access-98zzb") pod "e6ac4b62-a353-484c-ac04-77c77963cf9f" (UID: "e6ac4b62-a353-484c-ac04-77c77963cf9f"). InnerVolumeSpecName "kube-api-access-98zzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.654015 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf999f689-rksqm" event={"ID":"1553404c-fd63-4399-86bc-7b8550b8c372","Type":"ContainerStarted","Data":"7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830"} Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.656156 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-546d69f86c-qfggd" event={"ID":"e6ac4b62-a353-484c-ac04-77c77963cf9f","Type":"ContainerDied","Data":"75d8f3431302dc46f4771338bb89e2e201aacfa75604d8aa150a6fc647b11817"} Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.656201 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-qfggd" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.656806 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" event={"ID":"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c","Type":"ContainerDied","Data":"0ec6dd9e56d889d7697817b6bce31af365fc0b86c5726bf870445edc84c7aa84"} Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.656873 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-6rkzk" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.658105 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" event={"ID":"18a20547-c4f6-418d-ad4d-61116540d869","Type":"ContainerStarted","Data":"f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde"} Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.686607 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzpg8\" (UniqueName: \"kubernetes.io/projected/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-kube-api-access-xzpg8\") pod \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.686775 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-dns-svc\") pod \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.686853 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-config\") pod \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\" (UID: \"eb72eb23-93d7-4f24-aecf-e4f83ffaf24c\") " Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.687170 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6ac4b62-a353-484c-ac04-77c77963cf9f-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.687182 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98zzb\" (UniqueName: \"kubernetes.io/projected/e6ac4b62-a353-484c-ac04-77c77963cf9f-kube-api-access-98zzb\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.687673 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-config" (OuterVolumeSpecName: "config") pod "eb72eb23-93d7-4f24-aecf-e4f83ffaf24c" (UID: "eb72eb23-93d7-4f24-aecf-e4f83ffaf24c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.694137 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb72eb23-93d7-4f24-aecf-e4f83ffaf24c" (UID: "eb72eb23-93d7-4f24-aecf-e4f83ffaf24c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.695511 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-kube-api-access-xzpg8" (OuterVolumeSpecName: "kube-api-access-xzpg8") pod "eb72eb23-93d7-4f24-aecf-e4f83ffaf24c" (UID: "eb72eb23-93d7-4f24-aecf-e4f83ffaf24c"). InnerVolumeSpecName "kube-api-access-xzpg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.733201 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-qfggd"] Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.739433 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-qfggd"] Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.795801 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.795846 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.795867 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzpg8\" (UniqueName: \"kubernetes.io/projected/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c-kube-api-access-xzpg8\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.806166 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6ac4b62-a353-484c-ac04-77c77963cf9f" path="/var/lib/kubelet/pods/e6ac4b62-a353-484c-ac04-77c77963cf9f/volumes" Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.994609 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-6rkzk"] Oct 05 07:10:50 crc kubenswrapper[4935]: I1005 07:10:50.999502 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-6rkzk"] Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.067549 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.105505 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.114898 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.127285 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.134052 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-px6js"] Oct 05 07:10:51 crc kubenswrapper[4935]: W1005 07:10:51.135683 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0943221_d85f_4adb_a676_ffb7e7406eae.slice/crio-0b29c4e44ab55818c0973083601100ca8f55ace514463250bc75a11b71f20891 WatchSource:0}: Error finding container 0b29c4e44ab55818c0973083601100ca8f55ace514463250bc75a11b71f20891: Status 404 returned error can't find the container with id 0b29c4e44ab55818c0973083601100ca8f55ace514463250bc75a11b71f20891 Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.199840 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pl4nq"] Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.283483 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:10:51 crc kubenswrapper[4935]: W1005 07:10:51.364873 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaac8194b_9784_4eed_8c35_fb65a584b525.slice/crio-3a134aaab30465b85e8f3f839977f757b39b55e191ba8e33e47ccc0612c14aea WatchSource:0}: Error finding container 3a134aaab30465b85e8f3f839977f757b39b55e191ba8e33e47ccc0612c14aea: Status 404 returned error can't find the container with id 3a134aaab30465b85e8f3f839977f757b39b55e191ba8e33e47ccc0612c14aea Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.667518 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b82213d6-93fa-48fc-808b-36f902cce15c","Type":"ContainerStarted","Data":"325c92c64a0778df03cd0010f03519dc08238f80cfa11b3b7fbcc352581236cb"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.668984 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dcf22e2-2acd-4b49-a2bc-813a50f89409","Type":"ContainerStarted","Data":"7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.670928 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c0943221-d85f-4adb-a676-ffb7e7406eae","Type":"ContainerStarted","Data":"0b29c4e44ab55818c0973083601100ca8f55ace514463250bc75a11b71f20891"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.678404 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f27b0edc-d2a4-418e-8c62-82f1287cc1b4","Type":"ContainerStarted","Data":"7799e4cd4f8cf427bbdb16723cf47a8cede7eaf5eebd1674cb4e809090bd4dfc"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.680215 4935 generic.go:334] "Generic (PLEG): container finished" podID="18a20547-c4f6-418d-ad4d-61116540d869" containerID="f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde" exitCode=0 Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.680251 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" event={"ID":"18a20547-c4f6-418d-ad4d-61116540d869","Type":"ContainerDied","Data":"f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.685468 4935 generic.go:334] "Generic (PLEG): container finished" podID="1553404c-fd63-4399-86bc-7b8550b8c372" containerID="7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830" exitCode=0 Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.685679 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf999f689-rksqm" event={"ID":"1553404c-fd63-4399-86bc-7b8550b8c372","Type":"ContainerDied","Data":"7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.694211 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-px6js" event={"ID":"971ba03f-bdf3-4788-a28d-76ae1c5c658c","Type":"ContainerStarted","Data":"5fcd8d9784ce4232eba94e1b2dafd37ef5063910d6bc79f00d465760e9da532e"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.697209 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7bedc55a-2396-4249-ac54-a753a67dca72","Type":"ContainerStarted","Data":"857b569be287a3227232d02e565e22a5e29684e717b08f7787146b2c47045af3"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.698641 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"af041b3a-6ed3-4c8d-836b-a9b69f95360b","Type":"ContainerStarted","Data":"8496b167da6a6e05fbdffbb870edca983c3b713a036855854905138de03c1df0"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.700415 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"53fe24e1-5efb-4422-ade0-16e315b7e328","Type":"ContainerStarted","Data":"9c18cae5b21eea72b6ce894af8297def856aa4211aff70c789ca6375a34a3330"} Oct 05 07:10:51 crc kubenswrapper[4935]: I1005 07:10:51.703245 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pl4nq" event={"ID":"aac8194b-9784-4eed-8c35-fb65a584b525","Type":"ContainerStarted","Data":"3a134aaab30465b85e8f3f839977f757b39b55e191ba8e33e47ccc0612c14aea"} Oct 05 07:10:52 crc kubenswrapper[4935]: I1005 07:10:52.192809 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:10:52 crc kubenswrapper[4935]: I1005 07:10:52.720539 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0d984f8-20c4-4b30-b767-c6106aa279cf","Type":"ContainerStarted","Data":"8021a795afeb72e1c25c850f463e20c2b39ebf74611e8d459eb501f0960c8e2d"} Oct 05 07:10:52 crc kubenswrapper[4935]: I1005 07:10:52.788876 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb72eb23-93d7-4f24-aecf-e4f83ffaf24c" path="/var/lib/kubelet/pods/eb72eb23-93d7-4f24-aecf-e4f83ffaf24c/volumes" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.616123 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-l5t6l"] Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.617196 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.621403 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.626800 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-l5t6l"] Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.740300 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovn-rundir\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.740344 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-combined-ca-bundle\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.740382 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thh58\" (UniqueName: \"kubernetes.io/projected/ca591a2f-54ed-41b1-aa10-be879b218844-kube-api-access-thh58\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.740411 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovs-rundir\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.740437 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.740864 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca591a2f-54ed-41b1-aa10-be879b218844-config\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.842344 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca591a2f-54ed-41b1-aa10-be879b218844-config\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.842436 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovn-rundir\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.842460 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-combined-ca-bundle\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.842484 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thh58\" (UniqueName: \"kubernetes.io/projected/ca591a2f-54ed-41b1-aa10-be879b218844-kube-api-access-thh58\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.842506 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovs-rundir\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.842537 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.843689 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovs-rundir\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.843818 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovn-rundir\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.844497 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca591a2f-54ed-41b1-aa10-be879b218844-config\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.847412 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.851553 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-combined-ca-bundle\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.857666 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thh58\" (UniqueName: \"kubernetes.io/projected/ca591a2f-54ed-41b1-aa10-be879b218844-kube-api-access-thh58\") pod \"ovn-controller-metrics-l5t6l\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.946733 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:10:53 crc kubenswrapper[4935]: I1005 07:10:53.975656 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bf999f689-rksqm"] Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.004450 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-h2k56"] Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.007278 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.010648 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.013846 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-h2k56"] Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.046674 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-config\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.046853 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28wg5\" (UniqueName: \"kubernetes.io/projected/48d59f05-b045-417c-b42b-1f39adfaf0c6-kube-api-access-28wg5\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.047453 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-dns-svc\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.047641 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.108129 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-pgq4m"] Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.135430 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-ts2v6"] Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.137170 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.140087 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.148858 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.148949 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-config\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.148996 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28wg5\" (UniqueName: \"kubernetes.io/projected/48d59f05-b045-417c-b42b-1f39adfaf0c6-kube-api-access-28wg5\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.149071 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-dns-svc\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.150131 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-dns-svc\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.150214 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.150519 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-config\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.165119 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-ts2v6"] Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.176195 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28wg5\" (UniqueName: \"kubernetes.io/projected/48d59f05-b045-417c-b42b-1f39adfaf0c6-kube-api-access-28wg5\") pod \"dnsmasq-dns-6b8749979c-h2k56\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.250986 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-config\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.251074 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-nb\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.251108 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-sb\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.251167 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks5cq\" (UniqueName: \"kubernetes.io/projected/96e919ab-9dbf-4f54-9d45-8d52d59a6861-kube-api-access-ks5cq\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.251229 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-dns-svc\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.330793 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.352581 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-dns-svc\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.352641 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-config\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.352673 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-nb\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.352698 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-sb\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.352789 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks5cq\" (UniqueName: \"kubernetes.io/projected/96e919ab-9dbf-4f54-9d45-8d52d59a6861-kube-api-access-ks5cq\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.353619 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-nb\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.353634 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-sb\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.354070 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-config\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.354508 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-dns-svc\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.368596 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks5cq\" (UniqueName: \"kubernetes.io/projected/96e919ab-9dbf-4f54-9d45-8d52d59a6861-kube-api-access-ks5cq\") pod \"dnsmasq-dns-8454ffc489-ts2v6\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:54 crc kubenswrapper[4935]: I1005 07:10:54.471297 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:57 crc kubenswrapper[4935]: I1005 07:10:57.755565 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" event={"ID":"18a20547-c4f6-418d-ad4d-61116540d869","Type":"ContainerStarted","Data":"de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e"} Oct 05 07:10:57 crc kubenswrapper[4935]: I1005 07:10:57.756092 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:57 crc kubenswrapper[4935]: I1005 07:10:57.755741 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" podUID="18a20547-c4f6-418d-ad4d-61116540d869" containerName="dnsmasq-dns" containerID="cri-o://de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e" gracePeriod=10 Oct 05 07:10:57 crc kubenswrapper[4935]: I1005 07:10:57.775007 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" podStartSLOduration=8.38734939 podStartE2EDuration="21.774967761s" podCreationTimestamp="2025-10-05 07:10:36 +0000 UTC" firstStartedPulling="2025-10-05 07:10:37.001110882 +0000 UTC m=+1070.883737342" lastFinishedPulling="2025-10-05 07:10:50.388729253 +0000 UTC m=+1084.271355713" observedRunningTime="2025-10-05 07:10:57.769813244 +0000 UTC m=+1091.652439704" watchObservedRunningTime="2025-10-05 07:10:57.774967761 +0000 UTC m=+1091.657594221" Oct 05 07:10:57 crc kubenswrapper[4935]: I1005 07:10:57.794439 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-l5t6l"] Oct 05 07:10:57 crc kubenswrapper[4935]: W1005 07:10:57.850010 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca591a2f_54ed_41b1_aa10_be879b218844.slice/crio-d6edad8ea6f25d5359cf33edc983b72f7e48db74dfbbad4954ae2e4a2451ad55 WatchSource:0}: Error finding container d6edad8ea6f25d5359cf33edc983b72f7e48db74dfbbad4954ae2e4a2451ad55: Status 404 returned error can't find the container with id d6edad8ea6f25d5359cf33edc983b72f7e48db74dfbbad4954ae2e4a2451ad55 Oct 05 07:10:57 crc kubenswrapper[4935]: I1005 07:10:57.874212 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-ts2v6"] Oct 05 07:10:57 crc kubenswrapper[4935]: I1005 07:10:57.879609 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-h2k56"] Oct 05 07:10:57 crc kubenswrapper[4935]: W1005 07:10:57.884046 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96e919ab_9dbf_4f54_9d45_8d52d59a6861.slice/crio-fd5f9eb2b393d75edae9b848c6d923a8d451fbc0891cf9d3cca72fbf01fb4565 WatchSource:0}: Error finding container fd5f9eb2b393d75edae9b848c6d923a8d451fbc0891cf9d3cca72fbf01fb4565: Status 404 returned error can't find the container with id fd5f9eb2b393d75edae9b848c6d923a8d451fbc0891cf9d3cca72fbf01fb4565 Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.182587 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.245655 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-dns-svc\") pod \"18a20547-c4f6-418d-ad4d-61116540d869\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.245960 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgqk9\" (UniqueName: \"kubernetes.io/projected/18a20547-c4f6-418d-ad4d-61116540d869-kube-api-access-pgqk9\") pod \"18a20547-c4f6-418d-ad4d-61116540d869\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.246076 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-config\") pod \"18a20547-c4f6-418d-ad4d-61116540d869\" (UID: \"18a20547-c4f6-418d-ad4d-61116540d869\") " Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.256096 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a20547-c4f6-418d-ad4d-61116540d869-kube-api-access-pgqk9" (OuterVolumeSpecName: "kube-api-access-pgqk9") pod "18a20547-c4f6-418d-ad4d-61116540d869" (UID: "18a20547-c4f6-418d-ad4d-61116540d869"). InnerVolumeSpecName "kube-api-access-pgqk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.348682 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgqk9\" (UniqueName: \"kubernetes.io/projected/18a20547-c4f6-418d-ad4d-61116540d869-kube-api-access-pgqk9\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.418581 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "18a20547-c4f6-418d-ad4d-61116540d869" (UID: "18a20547-c4f6-418d-ad4d-61116540d869"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.422321 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-config" (OuterVolumeSpecName: "config") pod "18a20547-c4f6-418d-ad4d-61116540d869" (UID: "18a20547-c4f6-418d-ad4d-61116540d869"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.449994 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.450019 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18a20547-c4f6-418d-ad4d-61116540d869-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.766271 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf999f689-rksqm" event={"ID":"1553404c-fd63-4399-86bc-7b8550b8c372","Type":"ContainerStarted","Data":"93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.766386 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.766451 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bf999f689-rksqm" podUID="1553404c-fd63-4399-86bc-7b8550b8c372" containerName="dnsmasq-dns" containerID="cri-o://93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a" gracePeriod=10 Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.767805 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-px6js" event={"ID":"971ba03f-bdf3-4788-a28d-76ae1c5c658c","Type":"ContainerStarted","Data":"eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.767932 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-px6js" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.769490 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f27b0edc-d2a4-418e-8c62-82f1287cc1b4","Type":"ContainerStarted","Data":"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.771107 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-l5t6l" event={"ID":"ca591a2f-54ed-41b1-aa10-be879b218844","Type":"ContainerStarted","Data":"d6edad8ea6f25d5359cf33edc983b72f7e48db74dfbbad4954ae2e4a2451ad55"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.774130 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c0943221-d85f-4adb-a676-ffb7e7406eae","Type":"ContainerStarted","Data":"8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.774260 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.781143 4935 generic.go:334] "Generic (PLEG): container finished" podID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerID="91fe41dee07b7e68d2898b5a49350ca102e13d5605a862a8683544c80fc808a5" exitCode=0 Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.787659 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bf999f689-rksqm" podStartSLOduration=9.932203993 podStartE2EDuration="23.787645143s" podCreationTimestamp="2025-10-05 07:10:35 +0000 UTC" firstStartedPulling="2025-10-05 07:10:36.530704885 +0000 UTC m=+1070.413331345" lastFinishedPulling="2025-10-05 07:10:50.386146045 +0000 UTC m=+1084.268772495" observedRunningTime="2025-10-05 07:10:58.781272614 +0000 UTC m=+1092.663899074" watchObservedRunningTime="2025-10-05 07:10:58.787645143 +0000 UTC m=+1092.670271603" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.795648 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.795991 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pl4nq" event={"ID":"aac8194b-9784-4eed-8c35-fb65a584b525","Type":"ContainerStarted","Data":"2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.796010 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b82213d6-93fa-48fc-808b-36f902cce15c","Type":"ContainerStarted","Data":"fb6f5493ca2adedadd805d4c56c29f8cf8470abc87c240d1db6707f07565dd60"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.796020 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" event={"ID":"48d59f05-b045-417c-b42b-1f39adfaf0c6","Type":"ContainerDied","Data":"91fe41dee07b7e68d2898b5a49350ca102e13d5605a862a8683544c80fc808a5"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.796054 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" event={"ID":"48d59f05-b045-417c-b42b-1f39adfaf0c6","Type":"ContainerStarted","Data":"fc1f028969bea625d8b656b6e7c9f94b0e936a63ac20b2fa484c2a8b942149de"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.796065 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0d984f8-20c4-4b30-b767-c6106aa279cf","Type":"ContainerStarted","Data":"d3e5a40a1e26ef3caaba5145609516f5ff29e761ac28f571d0da1f626e989083"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.796075 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7bedc55a-2396-4249-ac54-a753a67dca72","Type":"ContainerStarted","Data":"9f89c3a737898aa877917fa322569786f07234b97e0640a587d32d6f00f43018"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.796086 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"af041b3a-6ed3-4c8d-836b-a9b69f95360b","Type":"ContainerStarted","Data":"7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.800882 4935 generic.go:334] "Generic (PLEG): container finished" podID="18a20547-c4f6-418d-ad4d-61116540d869" containerID="de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e" exitCode=0 Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.801170 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" event={"ID":"18a20547-c4f6-418d-ad4d-61116540d869","Type":"ContainerDied","Data":"de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.801252 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" event={"ID":"18a20547-c4f6-418d-ad4d-61116540d869","Type":"ContainerDied","Data":"2e91a7225d275977c5c162cdb7ce87691a935b3b8c57274eac9bc6617b74a17d"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.801293 4935 scope.go:117] "RemoveContainer" containerID="de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.811246 4935 generic.go:334] "Generic (PLEG): container finished" podID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" containerID="13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56" exitCode=0 Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.811325 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" event={"ID":"96e919ab-9dbf-4f54-9d45-8d52d59a6861","Type":"ContainerDied","Data":"13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.811364 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" event={"ID":"96e919ab-9dbf-4f54-9d45-8d52d59a6861","Type":"ContainerStarted","Data":"fd5f9eb2b393d75edae9b848c6d923a8d451fbc0891cf9d3cca72fbf01fb4565"} Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.825954 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-pgq4m" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.849308 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-px6js" podStartSLOduration=6.596824944 podStartE2EDuration="12.849293938s" podCreationTimestamp="2025-10-05 07:10:46 +0000 UTC" firstStartedPulling="2025-10-05 07:10:51.14486658 +0000 UTC m=+1085.027493040" lastFinishedPulling="2025-10-05 07:10:57.397335574 +0000 UTC m=+1091.279962034" observedRunningTime="2025-10-05 07:10:58.842913759 +0000 UTC m=+1092.725540219" watchObservedRunningTime="2025-10-05 07:10:58.849293938 +0000 UTC m=+1092.731920398" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.871391 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.028254003 podStartE2EDuration="16.871375724s" podCreationTimestamp="2025-10-05 07:10:42 +0000 UTC" firstStartedPulling="2025-10-05 07:10:51.141983314 +0000 UTC m=+1085.024609774" lastFinishedPulling="2025-10-05 07:10:57.985105035 +0000 UTC m=+1091.867731495" observedRunningTime="2025-10-05 07:10:58.857795773 +0000 UTC m=+1092.740422223" watchObservedRunningTime="2025-10-05 07:10:58.871375724 +0000 UTC m=+1092.754002184" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.885998 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.087721735 podStartE2EDuration="18.885977611s" podCreationTimestamp="2025-10-05 07:10:40 +0000 UTC" firstStartedPulling="2025-10-05 07:10:51.102134696 +0000 UTC m=+1084.984761156" lastFinishedPulling="2025-10-05 07:10:56.900390572 +0000 UTC m=+1090.783017032" observedRunningTime="2025-10-05 07:10:58.881904693 +0000 UTC m=+1092.764531153" watchObservedRunningTime="2025-10-05 07:10:58.885977611 +0000 UTC m=+1092.768604071" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.887950 4935 scope.go:117] "RemoveContainer" containerID="f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.954786 4935 scope.go:117] "RemoveContainer" containerID="de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e" Oct 05 07:10:58 crc kubenswrapper[4935]: E1005 07:10:58.957394 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e\": container with ID starting with de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e not found: ID does not exist" containerID="de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.957440 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e"} err="failed to get container status \"de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e\": rpc error: code = NotFound desc = could not find container \"de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e\": container with ID starting with de492bd6bfcb10821114cc9f3a6fdbf384125ebdc8f6128ba57386a09148a99e not found: ID does not exist" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.957471 4935 scope.go:117] "RemoveContainer" containerID="f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde" Oct 05 07:10:58 crc kubenswrapper[4935]: E1005 07:10:58.957981 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde\": container with ID starting with f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde not found: ID does not exist" containerID="f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.958010 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde"} err="failed to get container status \"f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde\": rpc error: code = NotFound desc = could not find container \"f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde\": container with ID starting with f57c2e0360853fb952a8eb39f96eeecc547d662b49e75bb8ffd15fcf00cd8bde not found: ID does not exist" Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.975405 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-pgq4m"] Oct 05 07:10:58 crc kubenswrapper[4935]: I1005 07:10:58.986454 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-pgq4m"] Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.362799 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.492727 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-dns-svc\") pod \"1553404c-fd63-4399-86bc-7b8550b8c372\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.492976 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-config\") pod \"1553404c-fd63-4399-86bc-7b8550b8c372\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.493049 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6qzb\" (UniqueName: \"kubernetes.io/projected/1553404c-fd63-4399-86bc-7b8550b8c372-kube-api-access-f6qzb\") pod \"1553404c-fd63-4399-86bc-7b8550b8c372\" (UID: \"1553404c-fd63-4399-86bc-7b8550b8c372\") " Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.499271 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1553404c-fd63-4399-86bc-7b8550b8c372-kube-api-access-f6qzb" (OuterVolumeSpecName: "kube-api-access-f6qzb") pod "1553404c-fd63-4399-86bc-7b8550b8c372" (UID: "1553404c-fd63-4399-86bc-7b8550b8c372"). InnerVolumeSpecName "kube-api-access-f6qzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.540401 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1553404c-fd63-4399-86bc-7b8550b8c372" (UID: "1553404c-fd63-4399-86bc-7b8550b8c372"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.562816 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-config" (OuterVolumeSpecName: "config") pod "1553404c-fd63-4399-86bc-7b8550b8c372" (UID: "1553404c-fd63-4399-86bc-7b8550b8c372"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.594736 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.594768 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6qzb\" (UniqueName: \"kubernetes.io/projected/1553404c-fd63-4399-86bc-7b8550b8c372-kube-api-access-f6qzb\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.594780 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1553404c-fd63-4399-86bc-7b8550b8c372-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.821853 4935 generic.go:334] "Generic (PLEG): container finished" podID="aac8194b-9784-4eed-8c35-fb65a584b525" containerID="2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095" exitCode=0 Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.821938 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pl4nq" event={"ID":"aac8194b-9784-4eed-8c35-fb65a584b525","Type":"ContainerDied","Data":"2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095"} Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.825871 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" event={"ID":"96e919ab-9dbf-4f54-9d45-8d52d59a6861","Type":"ContainerStarted","Data":"6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7"} Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.826130 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.830015 4935 generic.go:334] "Generic (PLEG): container finished" podID="1553404c-fd63-4399-86bc-7b8550b8c372" containerID="93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a" exitCode=0 Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.830070 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf999f689-rksqm" event={"ID":"1553404c-fd63-4399-86bc-7b8550b8c372","Type":"ContainerDied","Data":"93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a"} Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.830130 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bf999f689-rksqm" event={"ID":"1553404c-fd63-4399-86bc-7b8550b8c372","Type":"ContainerDied","Data":"0a926d1b8e37c004def352a3ab316a544f5c30a6a52e5abd86dda8b89b04f61a"} Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.830150 4935 scope.go:117] "RemoveContainer" containerID="93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.830080 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bf999f689-rksqm" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.832276 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" event={"ID":"48d59f05-b045-417c-b42b-1f39adfaf0c6","Type":"ContainerStarted","Data":"e5879599159900be29a83c3315965c32d70507f3276728809e65d1f558ebf4ce"} Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.857156 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" podStartSLOduration=6.857139772 podStartE2EDuration="6.857139772s" podCreationTimestamp="2025-10-05 07:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:59.855286833 +0000 UTC m=+1093.737913293" watchObservedRunningTime="2025-10-05 07:10:59.857139772 +0000 UTC m=+1093.739766232" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.872160 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" podStartSLOduration=5.87214081 podStartE2EDuration="5.87214081s" podCreationTimestamp="2025-10-05 07:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:59.871594576 +0000 UTC m=+1093.754221056" watchObservedRunningTime="2025-10-05 07:10:59.87214081 +0000 UTC m=+1093.754767260" Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.891339 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bf999f689-rksqm"] Oct 05 07:10:59 crc kubenswrapper[4935]: I1005 07:10:59.897317 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bf999f689-rksqm"] Oct 05 07:11:00 crc kubenswrapper[4935]: I1005 07:11:00.787815 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1553404c-fd63-4399-86bc-7b8550b8c372" path="/var/lib/kubelet/pods/1553404c-fd63-4399-86bc-7b8550b8c372/volumes" Oct 05 07:11:00 crc kubenswrapper[4935]: I1005 07:11:00.788428 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a20547-c4f6-418d-ad4d-61116540d869" path="/var/lib/kubelet/pods/18a20547-c4f6-418d-ad4d-61116540d869/volumes" Oct 05 07:11:00 crc kubenswrapper[4935]: I1005 07:11:00.840234 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:11:00 crc kubenswrapper[4935]: I1005 07:11:00.925233 4935 scope.go:117] "RemoveContainer" containerID="7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830" Oct 05 07:11:00 crc kubenswrapper[4935]: I1005 07:11:00.985680 4935 scope.go:117] "RemoveContainer" containerID="93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a" Oct 05 07:11:00 crc kubenswrapper[4935]: E1005 07:11:00.986155 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a\": container with ID starting with 93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a not found: ID does not exist" containerID="93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a" Oct 05 07:11:00 crc kubenswrapper[4935]: I1005 07:11:00.986196 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a"} err="failed to get container status \"93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a\": rpc error: code = NotFound desc = could not find container \"93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a\": container with ID starting with 93bad8fa9f5804fbd3b9b09e9c7b27c4b6e6be8657fd689971e0f0d4fbbb472a not found: ID does not exist" Oct 05 07:11:00 crc kubenswrapper[4935]: I1005 07:11:00.986230 4935 scope.go:117] "RemoveContainer" containerID="7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830" Oct 05 07:11:00 crc kubenswrapper[4935]: E1005 07:11:00.986533 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830\": container with ID starting with 7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830 not found: ID does not exist" containerID="7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830" Oct 05 07:11:00 crc kubenswrapper[4935]: I1005 07:11:00.986569 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830"} err="failed to get container status \"7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830\": rpc error: code = NotFound desc = could not find container \"7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830\": container with ID starting with 7fbc06d6e7436fd52e20c58fb19d29a188bac0d3726d9753b4d14ea5c1fbb830 not found: ID does not exist" Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.849142 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f27b0edc-d2a4-418e-8c62-82f1287cc1b4","Type":"ContainerStarted","Data":"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f"} Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.852825 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pl4nq" event={"ID":"aac8194b-9784-4eed-8c35-fb65a584b525","Type":"ContainerStarted","Data":"ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8"} Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.852851 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pl4nq" event={"ID":"aac8194b-9784-4eed-8c35-fb65a584b525","Type":"ContainerStarted","Data":"81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e"} Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.853534 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.853560 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.855929 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0d984f8-20c4-4b30-b767-c6106aa279cf","Type":"ContainerStarted","Data":"453dce919ffa6f7b445b83173fc85bcadcbbe96d524dbe18eb42e012aa8b54c1"} Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.857425 4935 generic.go:334] "Generic (PLEG): container finished" podID="7bedc55a-2396-4249-ac54-a753a67dca72" containerID="9f89c3a737898aa877917fa322569786f07234b97e0640a587d32d6f00f43018" exitCode=0 Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.857458 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7bedc55a-2396-4249-ac54-a753a67dca72","Type":"ContainerDied","Data":"9f89c3a737898aa877917fa322569786f07234b97e0640a587d32d6f00f43018"} Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.859012 4935 generic.go:334] "Generic (PLEG): container finished" podID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" containerID="7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f" exitCode=0 Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.859044 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"af041b3a-6ed3-4c8d-836b-a9b69f95360b","Type":"ContainerDied","Data":"7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f"} Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.863059 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-l5t6l" event={"ID":"ca591a2f-54ed-41b1-aa10-be879b218844","Type":"ContainerStarted","Data":"279993f164ba8d7c189ff4bd349158ae14baabd0b7dea6ef9cb73e2c002dc3b6"} Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.888204 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.260157993 podStartE2EDuration="13.888189607s" podCreationTimestamp="2025-10-05 07:10:48 +0000 UTC" firstStartedPulling="2025-10-05 07:10:51.368477942 +0000 UTC m=+1085.251104402" lastFinishedPulling="2025-10-05 07:11:00.996509556 +0000 UTC m=+1094.879136016" observedRunningTime="2025-10-05 07:11:01.868359831 +0000 UTC m=+1095.750986291" watchObservedRunningTime="2025-10-05 07:11:01.888189607 +0000 UTC m=+1095.770816067" Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.915957 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-l5t6l" podStartSLOduration=5.750895678 podStartE2EDuration="8.915879922s" podCreationTimestamp="2025-10-05 07:10:53 +0000 UTC" firstStartedPulling="2025-10-05 07:10:57.876038872 +0000 UTC m=+1091.758665332" lastFinishedPulling="2025-10-05 07:11:01.041023116 +0000 UTC m=+1094.923649576" observedRunningTime="2025-10-05 07:11:01.887315254 +0000 UTC m=+1095.769941734" watchObservedRunningTime="2025-10-05 07:11:01.915879922 +0000 UTC m=+1095.798506402" Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.950783 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.143777542 podStartE2EDuration="16.950763557s" podCreationTimestamp="2025-10-05 07:10:45 +0000 UTC" firstStartedPulling="2025-10-05 07:10:52.215664254 +0000 UTC m=+1086.098290714" lastFinishedPulling="2025-10-05 07:11:01.022650269 +0000 UTC m=+1094.905276729" observedRunningTime="2025-10-05 07:11:01.940911476 +0000 UTC m=+1095.823537936" watchObservedRunningTime="2025-10-05 07:11:01.950763557 +0000 UTC m=+1095.833390017" Oct 05 07:11:01 crc kubenswrapper[4935]: I1005 07:11:01.952384 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-pl4nq" podStartSLOduration=10.019870254 podStartE2EDuration="15.95237081s" podCreationTimestamp="2025-10-05 07:10:46 +0000 UTC" firstStartedPulling="2025-10-05 07:10:51.367048794 +0000 UTC m=+1085.249675254" lastFinishedPulling="2025-10-05 07:10:57.29954935 +0000 UTC m=+1091.182175810" observedRunningTime="2025-10-05 07:11:01.925733263 +0000 UTC m=+1095.808359733" watchObservedRunningTime="2025-10-05 07:11:01.95237081 +0000 UTC m=+1095.834997270" Oct 05 07:11:02 crc kubenswrapper[4935]: I1005 07:11:02.871046 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7bedc55a-2396-4249-ac54-a753a67dca72","Type":"ContainerStarted","Data":"5985a42517f3f6b261d93f0ab627c53c5bc297b9e7699a332b44b1e66d30ab31"} Oct 05 07:11:02 crc kubenswrapper[4935]: I1005 07:11:02.874051 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"af041b3a-6ed3-4c8d-836b-a9b69f95360b","Type":"ContainerStarted","Data":"09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d"} Oct 05 07:11:02 crc kubenswrapper[4935]: I1005 07:11:02.891835 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=17.598466111 podStartE2EDuration="23.89181496s" podCreationTimestamp="2025-10-05 07:10:39 +0000 UTC" firstStartedPulling="2025-10-05 07:10:51.109143162 +0000 UTC m=+1084.991769622" lastFinishedPulling="2025-10-05 07:10:57.402492011 +0000 UTC m=+1091.285118471" observedRunningTime="2025-10-05 07:11:02.890031043 +0000 UTC m=+1096.772657513" watchObservedRunningTime="2025-10-05 07:11:02.89181496 +0000 UTC m=+1096.774441420" Oct 05 07:11:02 crc kubenswrapper[4935]: I1005 07:11:02.915759 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=19.151441813 podStartE2EDuration="25.915737834s" podCreationTimestamp="2025-10-05 07:10:37 +0000 UTC" firstStartedPulling="2025-10-05 07:10:51.122751073 +0000 UTC m=+1085.005377533" lastFinishedPulling="2025-10-05 07:10:57.887047084 +0000 UTC m=+1091.769673554" observedRunningTime="2025-10-05 07:11:02.912778086 +0000 UTC m=+1096.795404546" watchObservedRunningTime="2025-10-05 07:11:02.915737834 +0000 UTC m=+1096.798364294" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.332811 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.473918 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.526280 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-h2k56"] Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.692455 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.754845 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.790396 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.790438 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.829295 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.886539 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" podUID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerName="dnsmasq-dns" containerID="cri-o://e5879599159900be29a83c3315965c32d70507f3276728809e65d1f558ebf4ce" gracePeriod=10 Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.888092 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.929128 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 05 07:11:04 crc kubenswrapper[4935]: I1005 07:11:04.930926 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.312773 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:11:05 crc kubenswrapper[4935]: E1005 07:11:05.313123 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1553404c-fd63-4399-86bc-7b8550b8c372" containerName="init" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.313140 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1553404c-fd63-4399-86bc-7b8550b8c372" containerName="init" Oct 05 07:11:05 crc kubenswrapper[4935]: E1005 07:11:05.313155 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a20547-c4f6-418d-ad4d-61116540d869" containerName="dnsmasq-dns" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.313162 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a20547-c4f6-418d-ad4d-61116540d869" containerName="dnsmasq-dns" Oct 05 07:11:05 crc kubenswrapper[4935]: E1005 07:11:05.313175 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1553404c-fd63-4399-86bc-7b8550b8c372" containerName="dnsmasq-dns" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.313182 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1553404c-fd63-4399-86bc-7b8550b8c372" containerName="dnsmasq-dns" Oct 05 07:11:05 crc kubenswrapper[4935]: E1005 07:11:05.313193 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a20547-c4f6-418d-ad4d-61116540d869" containerName="init" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.313199 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a20547-c4f6-418d-ad4d-61116540d869" containerName="init" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.313333 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="1553404c-fd63-4399-86bc-7b8550b8c372" containerName="dnsmasq-dns" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.313347 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a20547-c4f6-418d-ad4d-61116540d869" containerName="dnsmasq-dns" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.314120 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.316660 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.316765 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-s98m8" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.316985 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.349773 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.385182 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.451361 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-config\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.451421 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-scripts\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.451465 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.451531 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.451557 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.451639 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp5d6\" (UniqueName: \"kubernetes.io/projected/7c412f4e-54aa-4d6c-9606-753f97474c3b-kube-api-access-sp5d6\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.451794 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.552724 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.552819 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-config\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.552840 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-scripts\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.552870 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.552918 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.552937 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.552988 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp5d6\" (UniqueName: \"kubernetes.io/projected/7c412f4e-54aa-4d6c-9606-753f97474c3b-kube-api-access-sp5d6\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.553620 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.553986 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-config\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.554529 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-scripts\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.563403 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.563654 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.564753 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.578447 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp5d6\" (UniqueName: \"kubernetes.io/projected/7c412f4e-54aa-4d6c-9606-753f97474c3b-kube-api-access-sp5d6\") pod \"ovn-northd-0\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.659679 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.791682 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.894240 4935 generic.go:334] "Generic (PLEG): container finished" podID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerID="e5879599159900be29a83c3315965c32d70507f3276728809e65d1f558ebf4ce" exitCode=0 Oct 05 07:11:05 crc kubenswrapper[4935]: I1005 07:11:05.894950 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" event={"ID":"48d59f05-b045-417c-b42b-1f39adfaf0c6","Type":"ContainerDied","Data":"e5879599159900be29a83c3315965c32d70507f3276728809e65d1f558ebf4ce"} Oct 05 07:11:06 crc kubenswrapper[4935]: I1005 07:11:06.097799 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:11:06 crc kubenswrapper[4935]: W1005 07:11:06.102605 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c412f4e_54aa_4d6c_9606_753f97474c3b.slice/crio-f4fa02e54a41f112ed3743bd55ccac348e9e982e4645e9e8085878490221d327 WatchSource:0}: Error finding container f4fa02e54a41f112ed3743bd55ccac348e9e982e4645e9e8085878490221d327: Status 404 returned error can't find the container with id f4fa02e54a41f112ed3743bd55ccac348e9e982e4645e9e8085878490221d327 Oct 05 07:11:06 crc kubenswrapper[4935]: I1005 07:11:06.901271 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7c412f4e-54aa-4d6c-9606-753f97474c3b","Type":"ContainerStarted","Data":"f4fa02e54a41f112ed3743bd55ccac348e9e982e4645e9e8085878490221d327"} Oct 05 07:11:09 crc kubenswrapper[4935]: I1005 07:11:09.074010 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 05 07:11:09 crc kubenswrapper[4935]: I1005 07:11:09.074343 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 05 07:11:09 crc kubenswrapper[4935]: I1005 07:11:09.331956 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" podUID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Oct 05 07:11:10 crc kubenswrapper[4935]: I1005 07:11:10.458668 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 05 07:11:10 crc kubenswrapper[4935]: I1005 07:11:10.458825 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 05 07:11:11 crc kubenswrapper[4935]: I1005 07:11:11.969978 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.031110 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="7bedc55a-2396-4249-ac54-a753a67dca72" containerName="galera" probeResult="failure" output=< Oct 05 07:11:12 crc kubenswrapper[4935]: wsrep_local_state_comment (Joined) differs from Synced Oct 05 07:11:12 crc kubenswrapper[4935]: > Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.705101 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-vvdbb"] Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.706463 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.724395 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-vvdbb"] Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.762127 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.813755 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-sb\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.813804 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-nb\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.813854 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-config\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.813944 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvqmz\" (UniqueName: \"kubernetes.io/projected/116b1bb7-6886-431a-9127-41e40a8b1fd4-kube-api-access-pvqmz\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.814002 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-dns-svc\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.914906 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-dns-svc\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.914988 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-sb\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.915014 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-nb\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.915044 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-config\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.915085 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvqmz\" (UniqueName: \"kubernetes.io/projected/116b1bb7-6886-431a-9127-41e40a8b1fd4-kube-api-access-pvqmz\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.916114 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-nb\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.916218 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-sb\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.916300 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-config\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.916645 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-dns-svc\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:12 crc kubenswrapper[4935]: I1005 07:11:12.935684 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvqmz\" (UniqueName: \"kubernetes.io/projected/116b1bb7-6886-431a-9127-41e40a8b1fd4-kube-api-access-pvqmz\") pod \"dnsmasq-dns-665d79f5c9-vvdbb\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.027690 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.483643 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-vvdbb"] Oct 05 07:11:13 crc kubenswrapper[4935]: W1005 07:11:13.487104 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod116b1bb7_6886_431a_9127_41e40a8b1fd4.slice/crio-cdc7af747c4ea3c69ccaf5c3d22e1dacb5833052866ebc1a58b30e2d77181de5 WatchSource:0}: Error finding container cdc7af747c4ea3c69ccaf5c3d22e1dacb5833052866ebc1a58b30e2d77181de5: Status 404 returned error can't find the container with id cdc7af747c4ea3c69ccaf5c3d22e1dacb5833052866ebc1a58b30e2d77181de5 Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.902655 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.909717 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.912163 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.912293 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-vr9nf" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.912454 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.913546 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.932341 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.933125 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-cache\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.933155 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-lock\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.933174 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.933247 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.933319 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr2vk\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-kube-api-access-gr2vk\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.952087 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" event={"ID":"116b1bb7-6886-431a-9127-41e40a8b1fd4","Type":"ContainerStarted","Data":"3e7174750dde9789e06e2dd8200779016b4bc5330ebc8df3b13f82263e262821"} Oct 05 07:11:13 crc kubenswrapper[4935]: I1005 07:11:13.952151 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" event={"ID":"116b1bb7-6886-431a-9127-41e40a8b1fd4","Type":"ContainerStarted","Data":"cdc7af747c4ea3c69ccaf5c3d22e1dacb5833052866ebc1a58b30e2d77181de5"} Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.034920 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr2vk\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-kube-api-access-gr2vk\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.035577 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-cache\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.035606 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-lock\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.035627 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.035698 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: E1005 07:11:14.035820 4935 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:11:14 crc kubenswrapper[4935]: E1005 07:11:14.035834 4935 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:11:14 crc kubenswrapper[4935]: E1005 07:11:14.035874 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift podName:10a5c868-ffd7-44de-8f47-feb4c0ce9121 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:14.535858667 +0000 UTC m=+1108.418485127 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift") pod "swift-storage-0" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121") : configmap "swift-ring-files" not found Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.036639 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-cache\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.036848 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-lock\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.036951 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.059760 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr2vk\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-kube-api-access-gr2vk\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.062246 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.109080 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.239761 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-ovsdbserver-nb\") pod \"48d59f05-b045-417c-b42b-1f39adfaf0c6\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.239873 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-dns-svc\") pod \"48d59f05-b045-417c-b42b-1f39adfaf0c6\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.239981 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28wg5\" (UniqueName: \"kubernetes.io/projected/48d59f05-b045-417c-b42b-1f39adfaf0c6-kube-api-access-28wg5\") pod \"48d59f05-b045-417c-b42b-1f39adfaf0c6\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.240064 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-config\") pod \"48d59f05-b045-417c-b42b-1f39adfaf0c6\" (UID: \"48d59f05-b045-417c-b42b-1f39adfaf0c6\") " Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.244243 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d59f05-b045-417c-b42b-1f39adfaf0c6-kube-api-access-28wg5" (OuterVolumeSpecName: "kube-api-access-28wg5") pod "48d59f05-b045-417c-b42b-1f39adfaf0c6" (UID: "48d59f05-b045-417c-b42b-1f39adfaf0c6"). InnerVolumeSpecName "kube-api-access-28wg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.275330 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-config" (OuterVolumeSpecName: "config") pod "48d59f05-b045-417c-b42b-1f39adfaf0c6" (UID: "48d59f05-b045-417c-b42b-1f39adfaf0c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.278341 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48d59f05-b045-417c-b42b-1f39adfaf0c6" (UID: "48d59f05-b045-417c-b42b-1f39adfaf0c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.278376 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "48d59f05-b045-417c-b42b-1f39adfaf0c6" (UID: "48d59f05-b045-417c-b42b-1f39adfaf0c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.289509 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.289596 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.289676 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.290485 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be9c0311e869ecdb94c82f2824221de74d838dab99b4132c5a117c8e772b2a41"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.290542 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://be9c0311e869ecdb94c82f2824221de74d838dab99b4132c5a117c8e772b2a41" gracePeriod=600 Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.342431 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.342492 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28wg5\" (UniqueName: \"kubernetes.io/projected/48d59f05-b045-417c-b42b-1f39adfaf0c6-kube-api-access-28wg5\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.342551 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.342775 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48d59f05-b045-417c-b42b-1f39adfaf0c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.546875 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:14 crc kubenswrapper[4935]: E1005 07:11:14.547054 4935 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:11:14 crc kubenswrapper[4935]: E1005 07:11:14.547824 4935 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:11:14 crc kubenswrapper[4935]: E1005 07:11:14.547949 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift podName:10a5c868-ffd7-44de-8f47-feb4c0ce9121 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:15.54792031 +0000 UTC m=+1109.430546770 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift") pod "swift-storage-0" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121") : configmap "swift-ring-files" not found Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.963479 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="be9c0311e869ecdb94c82f2824221de74d838dab99b4132c5a117c8e772b2a41" exitCode=0 Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.963650 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"be9c0311e869ecdb94c82f2824221de74d838dab99b4132c5a117c8e772b2a41"} Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.963707 4935 scope.go:117] "RemoveContainer" containerID="52b267e7f2c1f92355886ddbf4cef30a218042be6e418507f266c72f41da3c70" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.966914 4935 generic.go:334] "Generic (PLEG): container finished" podID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerID="3e7174750dde9789e06e2dd8200779016b4bc5330ebc8df3b13f82263e262821" exitCode=0 Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.966967 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" event={"ID":"116b1bb7-6886-431a-9127-41e40a8b1fd4","Type":"ContainerDied","Data":"3e7174750dde9789e06e2dd8200779016b4bc5330ebc8df3b13f82263e262821"} Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.971710 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" event={"ID":"48d59f05-b045-417c-b42b-1f39adfaf0c6","Type":"ContainerDied","Data":"fc1f028969bea625d8b656b6e7c9f94b0e936a63ac20b2fa484c2a8b942149de"} Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.971809 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8749979c-h2k56" Oct 05 07:11:14 crc kubenswrapper[4935]: I1005 07:11:14.990016 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.024387 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-h2k56"] Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.033526 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-h2k56"] Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.047612 4935 scope.go:117] "RemoveContainer" containerID="e5879599159900be29a83c3315965c32d70507f3276728809e65d1f558ebf4ce" Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.071720 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.114079 4935 scope.go:117] "RemoveContainer" containerID="91fe41dee07b7e68d2898b5a49350ca102e13d5605a862a8683544c80fc808a5" Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.564794 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:15 crc kubenswrapper[4935]: E1005 07:11:15.565017 4935 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:11:15 crc kubenswrapper[4935]: E1005 07:11:15.565258 4935 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:11:15 crc kubenswrapper[4935]: E1005 07:11:15.565310 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift podName:10a5c868-ffd7-44de-8f47-feb4c0ce9121 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:17.565294198 +0000 UTC m=+1111.447920658 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift") pod "swift-storage-0" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121") : configmap "swift-ring-files" not found Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.988907 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"7b6d2f4c73d4ef3b7b40cc3341e20f2c6b3b4fe3d7affc8f4b8588270c63e170"} Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.993088 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" event={"ID":"116b1bb7-6886-431a-9127-41e40a8b1fd4","Type":"ContainerStarted","Data":"cdae626b037a8bbd125e95275e0ed740f42fd9e0890150e0768a66703b004ff3"} Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.993294 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:15 crc kubenswrapper[4935]: I1005 07:11:15.998634 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7c412f4e-54aa-4d6c-9606-753f97474c3b","Type":"ContainerStarted","Data":"1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6"} Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.031142 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" podStartSLOduration=4.031112683 podStartE2EDuration="4.031112683s" podCreationTimestamp="2025-10-05 07:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:11:16.025253958 +0000 UTC m=+1109.907880418" watchObservedRunningTime="2025-10-05 07:11:16.031112683 +0000 UTC m=+1109.913739163" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.068119 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-bdqj5"] Oct 05 07:11:16 crc kubenswrapper[4935]: E1005 07:11:16.068534 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerName="dnsmasq-dns" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.068550 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerName="dnsmasq-dns" Oct 05 07:11:16 crc kubenswrapper[4935]: E1005 07:11:16.068564 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerName="init" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.068570 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerName="init" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.068765 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d59f05-b045-417c-b42b-1f39adfaf0c6" containerName="dnsmasq-dns" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.069340 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bdqj5" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.073645 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bdqj5"] Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.176965 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmshx\" (UniqueName: \"kubernetes.io/projected/a1c2d664-55ad-4316-849a-b05d5a439388-kube-api-access-cmshx\") pod \"glance-db-create-bdqj5\" (UID: \"a1c2d664-55ad-4316-849a-b05d5a439388\") " pod="openstack/glance-db-create-bdqj5" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.278089 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmshx\" (UniqueName: \"kubernetes.io/projected/a1c2d664-55ad-4316-849a-b05d5a439388-kube-api-access-cmshx\") pod \"glance-db-create-bdqj5\" (UID: \"a1c2d664-55ad-4316-849a-b05d5a439388\") " pod="openstack/glance-db-create-bdqj5" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.306591 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmshx\" (UniqueName: \"kubernetes.io/projected/a1c2d664-55ad-4316-849a-b05d5a439388-kube-api-access-cmshx\") pod \"glance-db-create-bdqj5\" (UID: \"a1c2d664-55ad-4316-849a-b05d5a439388\") " pod="openstack/glance-db-create-bdqj5" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.384080 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bdqj5" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.788696 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d59f05-b045-417c-b42b-1f39adfaf0c6" path="/var/lib/kubelet/pods/48d59f05-b045-417c-b42b-1f39adfaf0c6/volumes" Oct 05 07:11:16 crc kubenswrapper[4935]: I1005 07:11:16.849441 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bdqj5"] Oct 05 07:11:16 crc kubenswrapper[4935]: W1005 07:11:16.857805 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1c2d664_55ad_4316_849a_b05d5a439388.slice/crio-63fd29c39d820337b72c7bce3101518ca6b053bc801e7dbb517c8916995f7049 WatchSource:0}: Error finding container 63fd29c39d820337b72c7bce3101518ca6b053bc801e7dbb517c8916995f7049: Status 404 returned error can't find the container with id 63fd29c39d820337b72c7bce3101518ca6b053bc801e7dbb517c8916995f7049 Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.011786 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bdqj5" event={"ID":"a1c2d664-55ad-4316-849a-b05d5a439388","Type":"ContainerStarted","Data":"63fd29c39d820337b72c7bce3101518ca6b053bc801e7dbb517c8916995f7049"} Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.014392 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7c412f4e-54aa-4d6c-9606-753f97474c3b","Type":"ContainerStarted","Data":"71a780f032841e0e4778281485674a6786913a58eb935c66f913a76d256ba68d"} Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.035057 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.463406455 podStartE2EDuration="12.035041114s" podCreationTimestamp="2025-10-05 07:11:05 +0000 UTC" firstStartedPulling="2025-10-05 07:11:06.104881679 +0000 UTC m=+1099.987508129" lastFinishedPulling="2025-10-05 07:11:15.676516328 +0000 UTC m=+1109.559142788" observedRunningTime="2025-10-05 07:11:17.03190458 +0000 UTC m=+1110.914531050" watchObservedRunningTime="2025-10-05 07:11:17.035041114 +0000 UTC m=+1110.917667574" Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.599721 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:17 crc kubenswrapper[4935]: E1005 07:11:17.599992 4935 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:11:17 crc kubenswrapper[4935]: E1005 07:11:17.600022 4935 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:11:17 crc kubenswrapper[4935]: E1005 07:11:17.600084 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift podName:10a5c868-ffd7-44de-8f47-feb4c0ce9121 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:21.600063192 +0000 UTC m=+1115.482689652 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift") pod "swift-storage-0" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121") : configmap "swift-ring-files" not found Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.819188 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-d9xn9"] Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.820222 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.822091 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.822090 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.825383 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 05 07:11:17 crc kubenswrapper[4935]: I1005 07:11:17.839210 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-d9xn9"] Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.007130 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/810c5536-9646-48ee-a6ac-04381b8c0cd5-etc-swift\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.007199 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-combined-ca-bundle\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.007381 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-ring-data-devices\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.007417 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-scripts\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.007578 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-dispersionconf\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.007689 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flbdx\" (UniqueName: \"kubernetes.io/projected/810c5536-9646-48ee-a6ac-04381b8c0cd5-kube-api-access-flbdx\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.007740 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-swiftconf\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.022351 4935 generic.go:334] "Generic (PLEG): container finished" podID="a1c2d664-55ad-4316-849a-b05d5a439388" containerID="c0b64cde633e9f03356159545948baedda001bcfee42e911ce7189993ae44f3a" exitCode=0 Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.022416 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bdqj5" event={"ID":"a1c2d664-55ad-4316-849a-b05d5a439388","Type":"ContainerDied","Data":"c0b64cde633e9f03356159545948baedda001bcfee42e911ce7189993ae44f3a"} Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.022717 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.109557 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-dispersionconf\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.109668 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flbdx\" (UniqueName: \"kubernetes.io/projected/810c5536-9646-48ee-a6ac-04381b8c0cd5-kube-api-access-flbdx\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.109712 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-swiftconf\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.109747 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/810c5536-9646-48ee-a6ac-04381b8c0cd5-etc-swift\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.109772 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-combined-ca-bundle\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.109865 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-ring-data-devices\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.109926 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-scripts\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.110346 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/810c5536-9646-48ee-a6ac-04381b8c0cd5-etc-swift\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.110916 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-ring-data-devices\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.110978 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-scripts\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.114775 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-swiftconf\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.114823 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-dispersionconf\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.115258 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-combined-ca-bundle\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.126217 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flbdx\" (UniqueName: \"kubernetes.io/projected/810c5536-9646-48ee-a6ac-04381b8c0cd5-kube-api-access-flbdx\") pod \"swift-ring-rebalance-d9xn9\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.147495 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:18 crc kubenswrapper[4935]: W1005 07:11:18.557625 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod810c5536_9646_48ee_a6ac_04381b8c0cd5.slice/crio-cc1536884e4d674b21ec12f88d54159ceb3ff0e2677f3765ba815037621fc03a WatchSource:0}: Error finding container cc1536884e4d674b21ec12f88d54159ceb3ff0e2677f3765ba815037621fc03a: Status 404 returned error can't find the container with id cc1536884e4d674b21ec12f88d54159ceb3ff0e2677f3765ba815037621fc03a Oct 05 07:11:18 crc kubenswrapper[4935]: I1005 07:11:18.557786 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-d9xn9"] Oct 05 07:11:19 crc kubenswrapper[4935]: I1005 07:11:19.032518 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d9xn9" event={"ID":"810c5536-9646-48ee-a6ac-04381b8c0cd5","Type":"ContainerStarted","Data":"cc1536884e4d674b21ec12f88d54159ceb3ff0e2677f3765ba815037621fc03a"} Oct 05 07:11:19 crc kubenswrapper[4935]: I1005 07:11:19.348494 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bdqj5" Oct 05 07:11:19 crc kubenswrapper[4935]: I1005 07:11:19.536401 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmshx\" (UniqueName: \"kubernetes.io/projected/a1c2d664-55ad-4316-849a-b05d5a439388-kube-api-access-cmshx\") pod \"a1c2d664-55ad-4316-849a-b05d5a439388\" (UID: \"a1c2d664-55ad-4316-849a-b05d5a439388\") " Oct 05 07:11:19 crc kubenswrapper[4935]: I1005 07:11:19.544515 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c2d664-55ad-4316-849a-b05d5a439388-kube-api-access-cmshx" (OuterVolumeSpecName: "kube-api-access-cmshx") pod "a1c2d664-55ad-4316-849a-b05d5a439388" (UID: "a1c2d664-55ad-4316-849a-b05d5a439388"). InnerVolumeSpecName "kube-api-access-cmshx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:19 crc kubenswrapper[4935]: I1005 07:11:19.638454 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmshx\" (UniqueName: \"kubernetes.io/projected/a1c2d664-55ad-4316-849a-b05d5a439388-kube-api-access-cmshx\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.041367 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bdqj5" event={"ID":"a1c2d664-55ad-4316-849a-b05d5a439388","Type":"ContainerDied","Data":"63fd29c39d820337b72c7bce3101518ca6b053bc801e7dbb517c8916995f7049"} Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.041629 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63fd29c39d820337b72c7bce3101518ca6b053bc801e7dbb517c8916995f7049" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.041422 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bdqj5" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.300246 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-dhmhx"] Oct 05 07:11:20 crc kubenswrapper[4935]: E1005 07:11:20.300664 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c2d664-55ad-4316-849a-b05d5a439388" containerName="mariadb-database-create" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.300685 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c2d664-55ad-4316-849a-b05d5a439388" containerName="mariadb-database-create" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.300832 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c2d664-55ad-4316-849a-b05d5a439388" containerName="mariadb-database-create" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.301344 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dhmhx" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.306182 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-dhmhx"] Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.453551 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2mqs\" (UniqueName: \"kubernetes.io/projected/ae9d561d-a53d-4b85-9232-776b137fb12c-kube-api-access-z2mqs\") pod \"keystone-db-create-dhmhx\" (UID: \"ae9d561d-a53d-4b85-9232-776b137fb12c\") " pod="openstack/keystone-db-create-dhmhx" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.512537 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.555784 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2mqs\" (UniqueName: \"kubernetes.io/projected/ae9d561d-a53d-4b85-9232-776b137fb12c-kube-api-access-z2mqs\") pod \"keystone-db-create-dhmhx\" (UID: \"ae9d561d-a53d-4b85-9232-776b137fb12c\") " pod="openstack/keystone-db-create-dhmhx" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.578011 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2mqs\" (UniqueName: \"kubernetes.io/projected/ae9d561d-a53d-4b85-9232-776b137fb12c-kube-api-access-z2mqs\") pod \"keystone-db-create-dhmhx\" (UID: \"ae9d561d-a53d-4b85-9232-776b137fb12c\") " pod="openstack/keystone-db-create-dhmhx" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.617943 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dhmhx" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.706809 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-z5h55"] Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.707867 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z5h55" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.712559 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-z5h55"] Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.861702 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wlp4\" (UniqueName: \"kubernetes.io/projected/42c5057a-3e6a-45fd-978c-6619b1a184ca-kube-api-access-8wlp4\") pod \"placement-db-create-z5h55\" (UID: \"42c5057a-3e6a-45fd-978c-6619b1a184ca\") " pod="openstack/placement-db-create-z5h55" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.963610 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wlp4\" (UniqueName: \"kubernetes.io/projected/42c5057a-3e6a-45fd-978c-6619b1a184ca-kube-api-access-8wlp4\") pod \"placement-db-create-z5h55\" (UID: \"42c5057a-3e6a-45fd-978c-6619b1a184ca\") " pod="openstack/placement-db-create-z5h55" Oct 05 07:11:20 crc kubenswrapper[4935]: I1005 07:11:20.979539 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wlp4\" (UniqueName: \"kubernetes.io/projected/42c5057a-3e6a-45fd-978c-6619b1a184ca-kube-api-access-8wlp4\") pod \"placement-db-create-z5h55\" (UID: \"42c5057a-3e6a-45fd-978c-6619b1a184ca\") " pod="openstack/placement-db-create-z5h55" Oct 05 07:11:21 crc kubenswrapper[4935]: I1005 07:11:21.030173 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z5h55" Oct 05 07:11:21 crc kubenswrapper[4935]: I1005 07:11:21.682107 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:21 crc kubenswrapper[4935]: E1005 07:11:21.682666 4935 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:11:21 crc kubenswrapper[4935]: E1005 07:11:21.682682 4935 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:11:21 crc kubenswrapper[4935]: E1005 07:11:21.682753 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift podName:10a5c868-ffd7-44de-8f47-feb4c0ce9121 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:29.682737589 +0000 UTC m=+1123.565364049 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift") pod "swift-storage-0" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121") : configmap "swift-ring-files" not found Oct 05 07:11:22 crc kubenswrapper[4935]: I1005 07:11:22.563799 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-dhmhx"] Oct 05 07:11:22 crc kubenswrapper[4935]: W1005 07:11:22.567065 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae9d561d_a53d_4b85_9232_776b137fb12c.slice/crio-51e7e89b6908d109f48127e9f4e6c0b356a2f030d5bcace19862adee19dd2e2b WatchSource:0}: Error finding container 51e7e89b6908d109f48127e9f4e6c0b356a2f030d5bcace19862adee19dd2e2b: Status 404 returned error can't find the container with id 51e7e89b6908d109f48127e9f4e6c0b356a2f030d5bcace19862adee19dd2e2b Oct 05 07:11:22 crc kubenswrapper[4935]: I1005 07:11:22.647099 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-z5h55"] Oct 05 07:11:22 crc kubenswrapper[4935]: W1005 07:11:22.647222 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42c5057a_3e6a_45fd_978c_6619b1a184ca.slice/crio-b904c3beada3dc999973a6342f16b3aea83f133629ffdc8ea81d800d338265e8 WatchSource:0}: Error finding container b904c3beada3dc999973a6342f16b3aea83f133629ffdc8ea81d800d338265e8: Status 404 returned error can't find the container with id b904c3beada3dc999973a6342f16b3aea83f133629ffdc8ea81d800d338265e8 Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.029383 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.067335 4935 generic.go:334] "Generic (PLEG): container finished" podID="42c5057a-3e6a-45fd-978c-6619b1a184ca" containerID="24c80e7ed845e73bf331fc87b91c537868571e5394f6a48ad9f2db5a21d684ab" exitCode=0 Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.067402 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z5h55" event={"ID":"42c5057a-3e6a-45fd-978c-6619b1a184ca","Type":"ContainerDied","Data":"24c80e7ed845e73bf331fc87b91c537868571e5394f6a48ad9f2db5a21d684ab"} Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.067431 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z5h55" event={"ID":"42c5057a-3e6a-45fd-978c-6619b1a184ca","Type":"ContainerStarted","Data":"b904c3beada3dc999973a6342f16b3aea83f133629ffdc8ea81d800d338265e8"} Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.069443 4935 generic.go:334] "Generic (PLEG): container finished" podID="ae9d561d-a53d-4b85-9232-776b137fb12c" containerID="d7c7e55c571cb0cfdbd49c39af5fe9080b3f370254ae163b6d61b41c9d8ee5cf" exitCode=0 Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.069486 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dhmhx" event={"ID":"ae9d561d-a53d-4b85-9232-776b137fb12c","Type":"ContainerDied","Data":"d7c7e55c571cb0cfdbd49c39af5fe9080b3f370254ae163b6d61b41c9d8ee5cf"} Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.069503 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dhmhx" event={"ID":"ae9d561d-a53d-4b85-9232-776b137fb12c","Type":"ContainerStarted","Data":"51e7e89b6908d109f48127e9f4e6c0b356a2f030d5bcace19862adee19dd2e2b"} Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.071417 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d9xn9" event={"ID":"810c5536-9646-48ee-a6ac-04381b8c0cd5","Type":"ContainerStarted","Data":"0f88e2c22830dd1c5d7712c2f78dedffa2c6264f0bb2ea8d8f08ba4a069716c1"} Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.116076 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-ts2v6"] Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.116381 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" podUID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" containerName="dnsmasq-dns" containerID="cri-o://6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7" gracePeriod=10 Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.142176 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-d9xn9" podStartSLOduration=2.6126135870000002 podStartE2EDuration="6.142156992s" podCreationTimestamp="2025-10-05 07:11:17 +0000 UTC" firstStartedPulling="2025-10-05 07:11:18.55981565 +0000 UTC m=+1112.442442110" lastFinishedPulling="2025-10-05 07:11:22.089359045 +0000 UTC m=+1115.971985515" observedRunningTime="2025-10-05 07:11:23.138379562 +0000 UTC m=+1117.021006062" watchObservedRunningTime="2025-10-05 07:11:23.142156992 +0000 UTC m=+1117.024783452" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.595733 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.623139 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks5cq\" (UniqueName: \"kubernetes.io/projected/96e919ab-9dbf-4f54-9d45-8d52d59a6861-kube-api-access-ks5cq\") pod \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.623228 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-dns-svc\") pod \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.623282 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-nb\") pod \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.623315 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-config\") pod \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.623345 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-sb\") pod \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\" (UID: \"96e919ab-9dbf-4f54-9d45-8d52d59a6861\") " Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.647238 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e919ab-9dbf-4f54-9d45-8d52d59a6861-kube-api-access-ks5cq" (OuterVolumeSpecName: "kube-api-access-ks5cq") pod "96e919ab-9dbf-4f54-9d45-8d52d59a6861" (UID: "96e919ab-9dbf-4f54-9d45-8d52d59a6861"). InnerVolumeSpecName "kube-api-access-ks5cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.668745 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "96e919ab-9dbf-4f54-9d45-8d52d59a6861" (UID: "96e919ab-9dbf-4f54-9d45-8d52d59a6861"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.678580 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-config" (OuterVolumeSpecName: "config") pod "96e919ab-9dbf-4f54-9d45-8d52d59a6861" (UID: "96e919ab-9dbf-4f54-9d45-8d52d59a6861"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.686642 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "96e919ab-9dbf-4f54-9d45-8d52d59a6861" (UID: "96e919ab-9dbf-4f54-9d45-8d52d59a6861"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.698495 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "96e919ab-9dbf-4f54-9d45-8d52d59a6861" (UID: "96e919ab-9dbf-4f54-9d45-8d52d59a6861"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.724332 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.724367 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.724376 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.724386 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks5cq\" (UniqueName: \"kubernetes.io/projected/96e919ab-9dbf-4f54-9d45-8d52d59a6861-kube-api-access-ks5cq\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4935]: I1005 07:11:23.724398 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96e919ab-9dbf-4f54-9d45-8d52d59a6861-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.085937 4935 generic.go:334] "Generic (PLEG): container finished" podID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerID="9c18cae5b21eea72b6ce894af8297def856aa4211aff70c789ca6375a34a3330" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.086233 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"53fe24e1-5efb-4422-ade0-16e315b7e328","Type":"ContainerDied","Data":"9c18cae5b21eea72b6ce894af8297def856aa4211aff70c789ca6375a34a3330"} Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.096705 4935 generic.go:334] "Generic (PLEG): container finished" podID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" containerID="7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.096770 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dcf22e2-2acd-4b49-a2bc-813a50f89409","Type":"ContainerDied","Data":"7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696"} Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.109702 4935 generic.go:334] "Generic (PLEG): container finished" podID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" containerID="6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.110312 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" event={"ID":"96e919ab-9dbf-4f54-9d45-8d52d59a6861","Type":"ContainerDied","Data":"6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7"} Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.110359 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" event={"ID":"96e919ab-9dbf-4f54-9d45-8d52d59a6861","Type":"ContainerDied","Data":"fd5f9eb2b393d75edae9b848c6d923a8d451fbc0891cf9d3cca72fbf01fb4565"} Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.110377 4935 scope.go:117] "RemoveContainer" containerID="6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.110525 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454ffc489-ts2v6" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.264414 4935 scope.go:117] "RemoveContainer" containerID="13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.279171 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-ts2v6"] Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.288431 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-ts2v6"] Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.311228 4935 scope.go:117] "RemoveContainer" containerID="6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7" Oct 05 07:11:24 crc kubenswrapper[4935]: E1005 07:11:24.312125 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7\": container with ID starting with 6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7 not found: ID does not exist" containerID="6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.312157 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7"} err="failed to get container status \"6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7\": rpc error: code = NotFound desc = could not find container \"6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7\": container with ID starting with 6e486193ad4d0b0910f8c8c30c310e00bc23c2e3cbc3b30509337d8f59e335a7 not found: ID does not exist" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.312176 4935 scope.go:117] "RemoveContainer" containerID="13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56" Oct 05 07:11:24 crc kubenswrapper[4935]: E1005 07:11:24.312978 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56\": container with ID starting with 13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56 not found: ID does not exist" containerID="13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.312998 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56"} err="failed to get container status \"13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56\": rpc error: code = NotFound desc = could not find container \"13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56\": container with ID starting with 13bcd040213d5a9ae5fce4ba772d0102f74edb948aefcd821cb54e46e7fd6f56 not found: ID does not exist" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.510564 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z5h55" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.554074 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dhmhx" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.636868 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wlp4\" (UniqueName: \"kubernetes.io/projected/42c5057a-3e6a-45fd-978c-6619b1a184ca-kube-api-access-8wlp4\") pod \"42c5057a-3e6a-45fd-978c-6619b1a184ca\" (UID: \"42c5057a-3e6a-45fd-978c-6619b1a184ca\") " Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.641124 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42c5057a-3e6a-45fd-978c-6619b1a184ca-kube-api-access-8wlp4" (OuterVolumeSpecName: "kube-api-access-8wlp4") pod "42c5057a-3e6a-45fd-978c-6619b1a184ca" (UID: "42c5057a-3e6a-45fd-978c-6619b1a184ca"). InnerVolumeSpecName "kube-api-access-8wlp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.737944 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2mqs\" (UniqueName: \"kubernetes.io/projected/ae9d561d-a53d-4b85-9232-776b137fb12c-kube-api-access-z2mqs\") pod \"ae9d561d-a53d-4b85-9232-776b137fb12c\" (UID: \"ae9d561d-a53d-4b85-9232-776b137fb12c\") " Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.738640 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wlp4\" (UniqueName: \"kubernetes.io/projected/42c5057a-3e6a-45fd-978c-6619b1a184ca-kube-api-access-8wlp4\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.743240 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae9d561d-a53d-4b85-9232-776b137fb12c-kube-api-access-z2mqs" (OuterVolumeSpecName: "kube-api-access-z2mqs") pod "ae9d561d-a53d-4b85-9232-776b137fb12c" (UID: "ae9d561d-a53d-4b85-9232-776b137fb12c"). InnerVolumeSpecName "kube-api-access-z2mqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.788366 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" path="/var/lib/kubelet/pods/96e919ab-9dbf-4f54-9d45-8d52d59a6861/volumes" Oct 05 07:11:24 crc kubenswrapper[4935]: I1005 07:11:24.840038 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2mqs\" (UniqueName: \"kubernetes.io/projected/ae9d561d-a53d-4b85-9232-776b137fb12c-kube-api-access-z2mqs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.121380 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dhmhx" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.121376 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dhmhx" event={"ID":"ae9d561d-a53d-4b85-9232-776b137fb12c","Type":"ContainerDied","Data":"51e7e89b6908d109f48127e9f4e6c0b356a2f030d5bcace19862adee19dd2e2b"} Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.122174 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51e7e89b6908d109f48127e9f4e6c0b356a2f030d5bcace19862adee19dd2e2b" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.123745 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"53fe24e1-5efb-4422-ade0-16e315b7e328","Type":"ContainerStarted","Data":"9e3ecf50ee2e44f330efce8b076a27f4149d1c011a741f1c565a07c5137e9563"} Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.123934 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.126719 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dcf22e2-2acd-4b49-a2bc-813a50f89409","Type":"ContainerStarted","Data":"cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a"} Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.127406 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.130650 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z5h55" event={"ID":"42c5057a-3e6a-45fd-978c-6619b1a184ca","Type":"ContainerDied","Data":"b904c3beada3dc999973a6342f16b3aea83f133629ffdc8ea81d800d338265e8"} Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.130681 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b904c3beada3dc999973a6342f16b3aea83f133629ffdc8ea81d800d338265e8" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.130693 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z5h55" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.145231 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.896689037 podStartE2EDuration="50.145219645s" podCreationTimestamp="2025-10-05 07:10:35 +0000 UTC" firstStartedPulling="2025-10-05 07:10:41.138709396 +0000 UTC m=+1075.021335856" lastFinishedPulling="2025-10-05 07:10:50.387240014 +0000 UTC m=+1084.269866464" observedRunningTime="2025-10-05 07:11:25.143099919 +0000 UTC m=+1119.025726379" watchObservedRunningTime="2025-10-05 07:11:25.145219645 +0000 UTC m=+1119.027846105" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.167088 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=47.812870182 podStartE2EDuration="49.167068244s" podCreationTimestamp="2025-10-05 07:10:36 +0000 UTC" firstStartedPulling="2025-10-05 07:10:49.033535135 +0000 UTC m=+1082.916161595" lastFinishedPulling="2025-10-05 07:10:50.387733197 +0000 UTC m=+1084.270359657" observedRunningTime="2025-10-05 07:11:25.164775324 +0000 UTC m=+1119.047401784" watchObservedRunningTime="2025-10-05 07:11:25.167068244 +0000 UTC m=+1119.049694704" Oct 05 07:11:25 crc kubenswrapper[4935]: I1005 07:11:25.715632 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.063664 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-e563-account-create-7ttqx"] Oct 05 07:11:26 crc kubenswrapper[4935]: E1005 07:11:26.063980 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c5057a-3e6a-45fd-978c-6619b1a184ca" containerName="mariadb-database-create" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.063992 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c5057a-3e6a-45fd-978c-6619b1a184ca" containerName="mariadb-database-create" Oct 05 07:11:26 crc kubenswrapper[4935]: E1005 07:11:26.064012 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" containerName="init" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.064019 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" containerName="init" Oct 05 07:11:26 crc kubenswrapper[4935]: E1005 07:11:26.064037 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" containerName="dnsmasq-dns" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.064045 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" containerName="dnsmasq-dns" Oct 05 07:11:26 crc kubenswrapper[4935]: E1005 07:11:26.064067 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9d561d-a53d-4b85-9232-776b137fb12c" containerName="mariadb-database-create" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.064074 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9d561d-a53d-4b85-9232-776b137fb12c" containerName="mariadb-database-create" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.064241 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9d561d-a53d-4b85-9232-776b137fb12c" containerName="mariadb-database-create" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.064254 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e919ab-9dbf-4f54-9d45-8d52d59a6861" containerName="dnsmasq-dns" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.064273 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="42c5057a-3e6a-45fd-978c-6619b1a184ca" containerName="mariadb-database-create" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.064759 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e563-account-create-7ttqx" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.069635 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.074931 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e563-account-create-7ttqx"] Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.174810 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cmcj\" (UniqueName: \"kubernetes.io/projected/61acbc8c-29cc-4afa-b3ce-34d65be5a32e-kube-api-access-7cmcj\") pod \"glance-e563-account-create-7ttqx\" (UID: \"61acbc8c-29cc-4afa-b3ce-34d65be5a32e\") " pod="openstack/glance-e563-account-create-7ttqx" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.277431 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cmcj\" (UniqueName: \"kubernetes.io/projected/61acbc8c-29cc-4afa-b3ce-34d65be5a32e-kube-api-access-7cmcj\") pod \"glance-e563-account-create-7ttqx\" (UID: \"61acbc8c-29cc-4afa-b3ce-34d65be5a32e\") " pod="openstack/glance-e563-account-create-7ttqx" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.306605 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cmcj\" (UniqueName: \"kubernetes.io/projected/61acbc8c-29cc-4afa-b3ce-34d65be5a32e-kube-api-access-7cmcj\") pod \"glance-e563-account-create-7ttqx\" (UID: \"61acbc8c-29cc-4afa-b3ce-34d65be5a32e\") " pod="openstack/glance-e563-account-create-7ttqx" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.382112 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e563-account-create-7ttqx" Oct 05 07:11:26 crc kubenswrapper[4935]: I1005 07:11:26.834309 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e563-account-create-7ttqx"] Oct 05 07:11:26 crc kubenswrapper[4935]: W1005 07:11:26.835779 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61acbc8c_29cc_4afa_b3ce_34d65be5a32e.slice/crio-17cc61ead2a9a9d8aee1af64457fc114a6ff8e22e7a5382a0ef74c4c5f2a9db9 WatchSource:0}: Error finding container 17cc61ead2a9a9d8aee1af64457fc114a6ff8e22e7a5382a0ef74c4c5f2a9db9: Status 404 returned error can't find the container with id 17cc61ead2a9a9d8aee1af64457fc114a6ff8e22e7a5382a0ef74c4c5f2a9db9 Oct 05 07:11:27 crc kubenswrapper[4935]: I1005 07:11:27.145530 4935 generic.go:334] "Generic (PLEG): container finished" podID="61acbc8c-29cc-4afa-b3ce-34d65be5a32e" containerID="183299868f8eadbbb5a725d9ebfc79c7434ded372ecb00a64b395cdd724042a7" exitCode=0 Oct 05 07:11:27 crc kubenswrapper[4935]: I1005 07:11:27.145689 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e563-account-create-7ttqx" event={"ID":"61acbc8c-29cc-4afa-b3ce-34d65be5a32e","Type":"ContainerDied","Data":"183299868f8eadbbb5a725d9ebfc79c7434ded372ecb00a64b395cdd724042a7"} Oct 05 07:11:27 crc kubenswrapper[4935]: I1005 07:11:27.145811 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e563-account-create-7ttqx" event={"ID":"61acbc8c-29cc-4afa-b3ce-34d65be5a32e","Type":"ContainerStarted","Data":"17cc61ead2a9a9d8aee1af64457fc114a6ff8e22e7a5382a0ef74c4c5f2a9db9"} Oct 05 07:11:28 crc kubenswrapper[4935]: I1005 07:11:28.439175 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e563-account-create-7ttqx" Oct 05 07:11:28 crc kubenswrapper[4935]: I1005 07:11:28.511177 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cmcj\" (UniqueName: \"kubernetes.io/projected/61acbc8c-29cc-4afa-b3ce-34d65be5a32e-kube-api-access-7cmcj\") pod \"61acbc8c-29cc-4afa-b3ce-34d65be5a32e\" (UID: \"61acbc8c-29cc-4afa-b3ce-34d65be5a32e\") " Oct 05 07:11:28 crc kubenswrapper[4935]: I1005 07:11:28.517374 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61acbc8c-29cc-4afa-b3ce-34d65be5a32e-kube-api-access-7cmcj" (OuterVolumeSpecName: "kube-api-access-7cmcj") pod "61acbc8c-29cc-4afa-b3ce-34d65be5a32e" (UID: "61acbc8c-29cc-4afa-b3ce-34d65be5a32e"). InnerVolumeSpecName "kube-api-access-7cmcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4935]: I1005 07:11:28.613098 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cmcj\" (UniqueName: \"kubernetes.io/projected/61acbc8c-29cc-4afa-b3ce-34d65be5a32e-kube-api-access-7cmcj\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4935]: I1005 07:11:29.160066 4935 generic.go:334] "Generic (PLEG): container finished" podID="810c5536-9646-48ee-a6ac-04381b8c0cd5" containerID="0f88e2c22830dd1c5d7712c2f78dedffa2c6264f0bb2ea8d8f08ba4a069716c1" exitCode=0 Oct 05 07:11:29 crc kubenswrapper[4935]: I1005 07:11:29.160174 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d9xn9" event={"ID":"810c5536-9646-48ee-a6ac-04381b8c0cd5","Type":"ContainerDied","Data":"0f88e2c22830dd1c5d7712c2f78dedffa2c6264f0bb2ea8d8f08ba4a069716c1"} Oct 05 07:11:29 crc kubenswrapper[4935]: I1005 07:11:29.161778 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e563-account-create-7ttqx" event={"ID":"61acbc8c-29cc-4afa-b3ce-34d65be5a32e","Type":"ContainerDied","Data":"17cc61ead2a9a9d8aee1af64457fc114a6ff8e22e7a5382a0ef74c4c5f2a9db9"} Oct 05 07:11:29 crc kubenswrapper[4935]: I1005 07:11:29.161818 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17cc61ead2a9a9d8aee1af64457fc114a6ff8e22e7a5382a0ef74c4c5f2a9db9" Oct 05 07:11:29 crc kubenswrapper[4935]: I1005 07:11:29.161880 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e563-account-create-7ttqx" Oct 05 07:11:29 crc kubenswrapper[4935]: I1005 07:11:29.729199 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:29 crc kubenswrapper[4935]: I1005 07:11:29.736636 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"swift-storage-0\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " pod="openstack/swift-storage-0" Oct 05 07:11:29 crc kubenswrapper[4935]: I1005 07:11:29.868674 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 07:11:30 crc kubenswrapper[4935]: W1005 07:11:30.428360 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a5c868_ffd7_44de_8f47_feb4c0ce9121.slice/crio-58c7a776ac2b258d5904745678ab9ccdba0ca364da2e464d53151208cc40c811 WatchSource:0}: Error finding container 58c7a776ac2b258d5904745678ab9ccdba0ca364da2e464d53151208cc40c811: Status 404 returned error can't find the container with id 58c7a776ac2b258d5904745678ab9ccdba0ca364da2e464d53151208cc40c811 Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.430519 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.447033 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.541884 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-combined-ca-bundle\") pod \"810c5536-9646-48ee-a6ac-04381b8c0cd5\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.542015 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-dispersionconf\") pod \"810c5536-9646-48ee-a6ac-04381b8c0cd5\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.542043 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-swiftconf\") pod \"810c5536-9646-48ee-a6ac-04381b8c0cd5\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.542170 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-ring-data-devices\") pod \"810c5536-9646-48ee-a6ac-04381b8c0cd5\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.542249 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-scripts\") pod \"810c5536-9646-48ee-a6ac-04381b8c0cd5\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.542286 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flbdx\" (UniqueName: \"kubernetes.io/projected/810c5536-9646-48ee-a6ac-04381b8c0cd5-kube-api-access-flbdx\") pod \"810c5536-9646-48ee-a6ac-04381b8c0cd5\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.542327 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/810c5536-9646-48ee-a6ac-04381b8c0cd5-etc-swift\") pod \"810c5536-9646-48ee-a6ac-04381b8c0cd5\" (UID: \"810c5536-9646-48ee-a6ac-04381b8c0cd5\") " Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.543370 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "810c5536-9646-48ee-a6ac-04381b8c0cd5" (UID: "810c5536-9646-48ee-a6ac-04381b8c0cd5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.543578 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810c5536-9646-48ee-a6ac-04381b8c0cd5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "810c5536-9646-48ee-a6ac-04381b8c0cd5" (UID: "810c5536-9646-48ee-a6ac-04381b8c0cd5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.543979 4935 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.544003 4935 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/810c5536-9646-48ee-a6ac-04381b8c0cd5-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.547841 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/810c5536-9646-48ee-a6ac-04381b8c0cd5-kube-api-access-flbdx" (OuterVolumeSpecName: "kube-api-access-flbdx") pod "810c5536-9646-48ee-a6ac-04381b8c0cd5" (UID: "810c5536-9646-48ee-a6ac-04381b8c0cd5"). InnerVolumeSpecName "kube-api-access-flbdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.550530 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "810c5536-9646-48ee-a6ac-04381b8c0cd5" (UID: "810c5536-9646-48ee-a6ac-04381b8c0cd5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.565155 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "810c5536-9646-48ee-a6ac-04381b8c0cd5" (UID: "810c5536-9646-48ee-a6ac-04381b8c0cd5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.565741 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-scripts" (OuterVolumeSpecName: "scripts") pod "810c5536-9646-48ee-a6ac-04381b8c0cd5" (UID: "810c5536-9646-48ee-a6ac-04381b8c0cd5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.570320 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "810c5536-9646-48ee-a6ac-04381b8c0cd5" (UID: "810c5536-9646-48ee-a6ac-04381b8c0cd5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.645344 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/810c5536-9646-48ee-a6ac-04381b8c0cd5-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.645631 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flbdx\" (UniqueName: \"kubernetes.io/projected/810c5536-9646-48ee-a6ac-04381b8c0cd5-kube-api-access-flbdx\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.645713 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.645778 4935 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.645835 4935 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/810c5536-9646-48ee-a6ac-04381b8c0cd5-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.750002 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-dd1f-account-create-tkrh9"] Oct 05 07:11:30 crc kubenswrapper[4935]: E1005 07:11:30.750406 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61acbc8c-29cc-4afa-b3ce-34d65be5a32e" containerName="mariadb-account-create" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.750424 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="61acbc8c-29cc-4afa-b3ce-34d65be5a32e" containerName="mariadb-account-create" Oct 05 07:11:30 crc kubenswrapper[4935]: E1005 07:11:30.750441 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810c5536-9646-48ee-a6ac-04381b8c0cd5" containerName="swift-ring-rebalance" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.750450 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="810c5536-9646-48ee-a6ac-04381b8c0cd5" containerName="swift-ring-rebalance" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.750654 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="61acbc8c-29cc-4afa-b3ce-34d65be5a32e" containerName="mariadb-account-create" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.750679 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="810c5536-9646-48ee-a6ac-04381b8c0cd5" containerName="swift-ring-rebalance" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.751320 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dd1f-account-create-tkrh9" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.755286 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.760744 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-dd1f-account-create-tkrh9"] Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.849433 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hv29\" (UniqueName: \"kubernetes.io/projected/6ef5a833-c6c9-43c7-ba5c-92b335d7ace7-kube-api-access-9hv29\") pod \"placement-dd1f-account-create-tkrh9\" (UID: \"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7\") " pod="openstack/placement-dd1f-account-create-tkrh9" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.950936 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hv29\" (UniqueName: \"kubernetes.io/projected/6ef5a833-c6c9-43c7-ba5c-92b335d7ace7-kube-api-access-9hv29\") pod \"placement-dd1f-account-create-tkrh9\" (UID: \"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7\") " pod="openstack/placement-dd1f-account-create-tkrh9" Oct 05 07:11:30 crc kubenswrapper[4935]: I1005 07:11:30.972300 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hv29\" (UniqueName: \"kubernetes.io/projected/6ef5a833-c6c9-43c7-ba5c-92b335d7ace7-kube-api-access-9hv29\") pod \"placement-dd1f-account-create-tkrh9\" (UID: \"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7\") " pod="openstack/placement-dd1f-account-create-tkrh9" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.091090 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dd1f-account-create-tkrh9" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.187105 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d9xn9" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.187599 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d9xn9" event={"ID":"810c5536-9646-48ee-a6ac-04381b8c0cd5","Type":"ContainerDied","Data":"cc1536884e4d674b21ec12f88d54159ceb3ff0e2677f3765ba815037621fc03a"} Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.187632 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc1536884e4d674b21ec12f88d54159ceb3ff0e2677f3765ba815037621fc03a" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.193057 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"58c7a776ac2b258d5904745678ab9ccdba0ca364da2e464d53151208cc40c811"} Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.302520 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-mbbgc"] Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.303747 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.307385 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g86zq" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.307457 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.311582 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mbbgc"] Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.458752 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-db-sync-config-data\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.458842 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-265rb\" (UniqueName: \"kubernetes.io/projected/2da110a9-7934-4fc0-88dc-8025c46413cb-kube-api-access-265rb\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.458884 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-config-data\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.459019 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-combined-ca-bundle\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.560172 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-combined-ca-bundle\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.560349 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-db-sync-config-data\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.560442 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-265rb\" (UniqueName: \"kubernetes.io/projected/2da110a9-7934-4fc0-88dc-8025c46413cb-kube-api-access-265rb\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.560470 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-config-data\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.565213 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-db-sync-config-data\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.565349 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-config-data\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.578717 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-combined-ca-bundle\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.596296 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-dd1f-account-create-tkrh9"] Oct 05 07:11:31 crc kubenswrapper[4935]: W1005 07:11:31.597359 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ef5a833_c6c9_43c7_ba5c_92b335d7ace7.slice/crio-769f2a7e37081f8c51ad5dbad7d0da42d24f38adcf3df5d47c172c33f55a17d6 WatchSource:0}: Error finding container 769f2a7e37081f8c51ad5dbad7d0da42d24f38adcf3df5d47c172c33f55a17d6: Status 404 returned error can't find the container with id 769f2a7e37081f8c51ad5dbad7d0da42d24f38adcf3df5d47c172c33f55a17d6 Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.597504 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-265rb\" (UniqueName: \"kubernetes.io/projected/2da110a9-7934-4fc0-88dc-8025c46413cb-kube-api-access-265rb\") pod \"glance-db-sync-mbbgc\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.627436 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mbbgc" Oct 05 07:11:31 crc kubenswrapper[4935]: I1005 07:11:31.931769 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mbbgc"] Oct 05 07:11:31 crc kubenswrapper[4935]: W1005 07:11:31.938405 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2da110a9_7934_4fc0_88dc_8025c46413cb.slice/crio-29a22f245ecdc558f6f9a60092d7247ab6ca51e68ea6ec5c3ef1f08b689587f4 WatchSource:0}: Error finding container 29a22f245ecdc558f6f9a60092d7247ab6ca51e68ea6ec5c3ef1f08b689587f4: Status 404 returned error can't find the container with id 29a22f245ecdc558f6f9a60092d7247ab6ca51e68ea6ec5c3ef1f08b689587f4 Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.208416 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dd1f-account-create-tkrh9" event={"ID":"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7","Type":"ContainerStarted","Data":"a8d4be873da7a3591175629c55ecf97d564e67d03648bff86a6511dce9218fce"} Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.209370 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dd1f-account-create-tkrh9" event={"ID":"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7","Type":"ContainerStarted","Data":"769f2a7e37081f8c51ad5dbad7d0da42d24f38adcf3df5d47c172c33f55a17d6"} Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.209427 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mbbgc" event={"ID":"2da110a9-7934-4fc0-88dc-8025c46413cb","Type":"ContainerStarted","Data":"29a22f245ecdc558f6f9a60092d7247ab6ca51e68ea6ec5c3ef1f08b689587f4"} Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.299289 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-px6js" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerName="ovn-controller" probeResult="failure" output=< Oct 05 07:11:32 crc kubenswrapper[4935]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 05 07:11:32 crc kubenswrapper[4935]: > Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.336080 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.336158 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.533827 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-px6js-config-jzrl8"] Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.534800 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.537275 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.544482 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-px6js-config-jzrl8"] Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.681858 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run-ovn\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.682275 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq69v\" (UniqueName: \"kubernetes.io/projected/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-kube-api-access-gq69v\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.682354 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-additional-scripts\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.682481 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-scripts\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.682572 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-log-ovn\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.682614 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.784315 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run-ovn\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.784375 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq69v\" (UniqueName: \"kubernetes.io/projected/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-kube-api-access-gq69v\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.784402 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-additional-scripts\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.784434 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-scripts\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.784462 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-log-ovn\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.784483 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.784717 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.784762 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-log-ovn\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.785421 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-additional-scripts\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.786581 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-scripts\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.786659 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run-ovn\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.806781 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq69v\" (UniqueName: \"kubernetes.io/projected/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-kube-api-access-gq69v\") pod \"ovn-controller-px6js-config-jzrl8\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:32 crc kubenswrapper[4935]: I1005 07:11:32.855726 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:33 crc kubenswrapper[4935]: I1005 07:11:33.221512 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"22ba6a3f7d0c659cde8f9b4b899df8e9b5b426c7a3f7e76b86ffc508196c5584"} Oct 05 07:11:33 crc kubenswrapper[4935]: I1005 07:11:33.221851 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"a461381cdfc3c2eee74eb1bdcf7d5ebfdb0fb05a83a09c83b9bcd518fb51b2d5"} Oct 05 07:11:33 crc kubenswrapper[4935]: I1005 07:11:33.224007 4935 generic.go:334] "Generic (PLEG): container finished" podID="6ef5a833-c6c9-43c7-ba5c-92b335d7ace7" containerID="a8d4be873da7a3591175629c55ecf97d564e67d03648bff86a6511dce9218fce" exitCode=0 Oct 05 07:11:33 crc kubenswrapper[4935]: I1005 07:11:33.224033 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dd1f-account-create-tkrh9" event={"ID":"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7","Type":"ContainerDied","Data":"a8d4be873da7a3591175629c55ecf97d564e67d03648bff86a6511dce9218fce"} Oct 05 07:11:33 crc kubenswrapper[4935]: I1005 07:11:33.346967 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-px6js-config-jzrl8"] Oct 05 07:11:33 crc kubenswrapper[4935]: W1005 07:11:33.356207 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5cc83e3_95d5_4ec2_8757_bb528e407fb7.slice/crio-d7ed36941ffd56156c812c9202edf85cd8af7aeceeac8622e4053f82dc28d2cc WatchSource:0}: Error finding container d7ed36941ffd56156c812c9202edf85cd8af7aeceeac8622e4053f82dc28d2cc: Status 404 returned error can't find the container with id d7ed36941ffd56156c812c9202edf85cd8af7aeceeac8622e4053f82dc28d2cc Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.234684 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"95b99cc19f9b362efc74b2612128d2a9a4247e4bfa40d9b7ca5e1c764c78c5c9"} Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.234944 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"5fb020c3045a74d5048d67e9e8f939b0d2f9348bab133b9b65f5cfb8ff321504"} Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.237944 4935 generic.go:334] "Generic (PLEG): container finished" podID="c5cc83e3-95d5-4ec2-8757-bb528e407fb7" containerID="acc87771f402f72e5094e197427a1e741957a406635fe722bbeda86770f046c0" exitCode=0 Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.238003 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-px6js-config-jzrl8" event={"ID":"c5cc83e3-95d5-4ec2-8757-bb528e407fb7","Type":"ContainerDied","Data":"acc87771f402f72e5094e197427a1e741957a406635fe722bbeda86770f046c0"} Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.238279 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-px6js-config-jzrl8" event={"ID":"c5cc83e3-95d5-4ec2-8757-bb528e407fb7","Type":"ContainerStarted","Data":"d7ed36941ffd56156c812c9202edf85cd8af7aeceeac8622e4053f82dc28d2cc"} Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.714153 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dd1f-account-create-tkrh9" Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.820019 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hv29\" (UniqueName: \"kubernetes.io/projected/6ef5a833-c6c9-43c7-ba5c-92b335d7ace7-kube-api-access-9hv29\") pod \"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7\" (UID: \"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7\") " Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.825764 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ef5a833-c6c9-43c7-ba5c-92b335d7ace7-kube-api-access-9hv29" (OuterVolumeSpecName: "kube-api-access-9hv29") pod "6ef5a833-c6c9-43c7-ba5c-92b335d7ace7" (UID: "6ef5a833-c6c9-43c7-ba5c-92b335d7ace7"). InnerVolumeSpecName "kube-api-access-9hv29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:34 crc kubenswrapper[4935]: I1005 07:11:34.921847 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hv29\" (UniqueName: \"kubernetes.io/projected/6ef5a833-c6c9-43c7-ba5c-92b335d7ace7-kube-api-access-9hv29\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.256344 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"cd237a014093f689d7df188672e570ebe94fb2b0f49e1d1a6c58473d64c3b740"} Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.256387 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"b18142ab07c004a98bff59b529be5e4c49e5ff54caf77526350e39ce8c20c39c"} Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.256398 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"cd51894ecb6651dbb869d73d13d130a6f2fb7d28b0f7980f406450f51c4449b9"} Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.260785 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dd1f-account-create-tkrh9" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.269026 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dd1f-account-create-tkrh9" event={"ID":"6ef5a833-c6c9-43c7-ba5c-92b335d7ace7","Type":"ContainerDied","Data":"769f2a7e37081f8c51ad5dbad7d0da42d24f38adcf3df5d47c172c33f55a17d6"} Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.269056 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="769f2a7e37081f8c51ad5dbad7d0da42d24f38adcf3df5d47c172c33f55a17d6" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.530526 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641269 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-scripts\") pod \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641366 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run\") pod \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641507 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run" (OuterVolumeSpecName: "var-run") pod "c5cc83e3-95d5-4ec2-8757-bb528e407fb7" (UID: "c5cc83e3-95d5-4ec2-8757-bb528e407fb7"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641591 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run-ovn\") pod \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641619 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-log-ovn\") pod \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641662 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c5cc83e3-95d5-4ec2-8757-bb528e407fb7" (UID: "c5cc83e3-95d5-4ec2-8757-bb528e407fb7"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641741 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-additional-scripts\") pod \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641713 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c5cc83e3-95d5-4ec2-8757-bb528e407fb7" (UID: "c5cc83e3-95d5-4ec2-8757-bb528e407fb7"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.641875 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq69v\" (UniqueName: \"kubernetes.io/projected/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-kube-api-access-gq69v\") pod \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\" (UID: \"c5cc83e3-95d5-4ec2-8757-bb528e407fb7\") " Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.642563 4935 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.642596 4935 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.642608 4935 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.642682 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-scripts" (OuterVolumeSpecName: "scripts") pod "c5cc83e3-95d5-4ec2-8757-bb528e407fb7" (UID: "c5cc83e3-95d5-4ec2-8757-bb528e407fb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.642943 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c5cc83e3-95d5-4ec2-8757-bb528e407fb7" (UID: "c5cc83e3-95d5-4ec2-8757-bb528e407fb7"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.648725 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-kube-api-access-gq69v" (OuterVolumeSpecName: "kube-api-access-gq69v") pod "c5cc83e3-95d5-4ec2-8757-bb528e407fb7" (UID: "c5cc83e3-95d5-4ec2-8757-bb528e407fb7"). InnerVolumeSpecName "kube-api-access-gq69v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.744386 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq69v\" (UniqueName: \"kubernetes.io/projected/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-kube-api-access-gq69v\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.744420 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:35 crc kubenswrapper[4935]: I1005 07:11:35.744430 4935 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c5cc83e3-95d5-4ec2-8757-bb528e407fb7-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:36 crc kubenswrapper[4935]: I1005 07:11:36.273565 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-px6js-config-jzrl8" event={"ID":"c5cc83e3-95d5-4ec2-8757-bb528e407fb7","Type":"ContainerDied","Data":"d7ed36941ffd56156c812c9202edf85cd8af7aeceeac8622e4053f82dc28d2cc"} Oct 05 07:11:36 crc kubenswrapper[4935]: I1005 07:11:36.273601 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7ed36941ffd56156c812c9202edf85cd8af7aeceeac8622e4053f82dc28d2cc" Oct 05 07:11:36 crc kubenswrapper[4935]: I1005 07:11:36.273633 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-px6js-config-jzrl8" Oct 05 07:11:36 crc kubenswrapper[4935]: I1005 07:11:36.280156 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"ba8102ea51bc30c759201a866a8431d1baaf2f6f474cc96b36529ea5a04451ec"} Oct 05 07:11:36 crc kubenswrapper[4935]: I1005 07:11:36.622729 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-px6js-config-jzrl8"] Oct 05 07:11:36 crc kubenswrapper[4935]: I1005 07:11:36.629556 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-px6js-config-jzrl8"] Oct 05 07:11:36 crc kubenswrapper[4935]: I1005 07:11:36.811135 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5cc83e3-95d5-4ec2-8757-bb528e407fb7" path="/var/lib/kubelet/pods/c5cc83e3-95d5-4ec2-8757-bb528e407fb7/volumes" Oct 05 07:11:37 crc kubenswrapper[4935]: I1005 07:11:37.069093 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:11:37 crc kubenswrapper[4935]: I1005 07:11:37.269036 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-px6js" Oct 05 07:11:37 crc kubenswrapper[4935]: I1005 07:11:37.959071 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.321204 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"8ae6a7f5c8eb347b93d521a8aa9df7e0b4edee491ce3dd6e4dd29fd1cdc12733"} Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.321326 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"5811d6a7453e1bdaa6a37fe17b4607303b6b13e3e3d1fe5e319d227f3a5e3efb"} Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.321342 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"f07908304e5257af7f03f59b8ba6637156f346dca17601807d91da7efcc445e9"} Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.321351 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"06e262448da43be1f59f8206d3ca149b097157a5612b8e5d368736aac9a62a57"} Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.904799 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-5z299"] Oct 05 07:11:38 crc kubenswrapper[4935]: E1005 07:11:38.906282 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cc83e3-95d5-4ec2-8757-bb528e407fb7" containerName="ovn-config" Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.906311 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cc83e3-95d5-4ec2-8757-bb528e407fb7" containerName="ovn-config" Oct 05 07:11:38 crc kubenswrapper[4935]: E1005 07:11:38.906353 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ef5a833-c6c9-43c7-ba5c-92b335d7ace7" containerName="mariadb-account-create" Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.906360 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ef5a833-c6c9-43c7-ba5c-92b335d7ace7" containerName="mariadb-account-create" Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.906763 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cc83e3-95d5-4ec2-8757-bb528e407fb7" containerName="ovn-config" Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.906794 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ef5a833-c6c9-43c7-ba5c-92b335d7ace7" containerName="mariadb-account-create" Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.907607 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5z299" Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.919928 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5z299"] Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.979313 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-kpphs"] Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.980994 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kpphs" Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.987415 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-kpphs"] Oct 05 07:11:38 crc kubenswrapper[4935]: I1005 07:11:38.998052 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4p9s\" (UniqueName: \"kubernetes.io/projected/c7a3c603-bb1a-4a34-80e3-4cae0ef2815b-kube-api-access-c4p9s\") pod \"cinder-db-create-5z299\" (UID: \"c7a3c603-bb1a-4a34-80e3-4cae0ef2815b\") " pod="openstack/cinder-db-create-5z299" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.099191 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4p9s\" (UniqueName: \"kubernetes.io/projected/c7a3c603-bb1a-4a34-80e3-4cae0ef2815b-kube-api-access-c4p9s\") pod \"cinder-db-create-5z299\" (UID: \"c7a3c603-bb1a-4a34-80e3-4cae0ef2815b\") " pod="openstack/cinder-db-create-5z299" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.099235 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/8a278b3a-993f-427f-9ab5-d189cbfb4a69-kube-api-access-z2kcb\") pod \"barbican-db-create-kpphs\" (UID: \"8a278b3a-993f-427f-9ab5-d189cbfb4a69\") " pod="openstack/barbican-db-create-kpphs" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.116130 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4p9s\" (UniqueName: \"kubernetes.io/projected/c7a3c603-bb1a-4a34-80e3-4cae0ef2815b-kube-api-access-c4p9s\") pod \"cinder-db-create-5z299\" (UID: \"c7a3c603-bb1a-4a34-80e3-4cae0ef2815b\") " pod="openstack/cinder-db-create-5z299" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.200959 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/8a278b3a-993f-427f-9ab5-d189cbfb4a69-kube-api-access-z2kcb\") pod \"barbican-db-create-kpphs\" (UID: \"8a278b3a-993f-427f-9ab5-d189cbfb4a69\") " pod="openstack/barbican-db-create-kpphs" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.219419 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/8a278b3a-993f-427f-9ab5-d189cbfb4a69-kube-api-access-z2kcb\") pod \"barbican-db-create-kpphs\" (UID: \"8a278b3a-993f-427f-9ab5-d189cbfb4a69\") " pod="openstack/barbican-db-create-kpphs" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.227536 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5z299" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.277974 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-b2q9j"] Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.279281 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b2q9j" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.294761 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kpphs" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.295608 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-b2q9j"] Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.404109 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dlnq\" (UniqueName: \"kubernetes.io/projected/0bf18d65-b3e6-41ef-a263-bd8642d272ed-kube-api-access-4dlnq\") pod \"neutron-db-create-b2q9j\" (UID: \"0bf18d65-b3e6-41ef-a263-bd8642d272ed\") " pod="openstack/neutron-db-create-b2q9j" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.505672 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dlnq\" (UniqueName: \"kubernetes.io/projected/0bf18d65-b3e6-41ef-a263-bd8642d272ed-kube-api-access-4dlnq\") pod \"neutron-db-create-b2q9j\" (UID: \"0bf18d65-b3e6-41ef-a263-bd8642d272ed\") " pod="openstack/neutron-db-create-b2q9j" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.534515 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dlnq\" (UniqueName: \"kubernetes.io/projected/0bf18d65-b3e6-41ef-a263-bd8642d272ed-kube-api-access-4dlnq\") pod \"neutron-db-create-b2q9j\" (UID: \"0bf18d65-b3e6-41ef-a263-bd8642d272ed\") " pod="openstack/neutron-db-create-b2q9j" Oct 05 07:11:39 crc kubenswrapper[4935]: I1005 07:11:39.598048 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b2q9j" Oct 05 07:11:40 crc kubenswrapper[4935]: I1005 07:11:40.407709 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5ce4-account-create-n7jk5"] Oct 05 07:11:40 crc kubenswrapper[4935]: I1005 07:11:40.408719 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ce4-account-create-n7jk5" Oct 05 07:11:40 crc kubenswrapper[4935]: I1005 07:11:40.411461 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 05 07:11:40 crc kubenswrapper[4935]: I1005 07:11:40.419939 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5ce4-account-create-n7jk5"] Oct 05 07:11:40 crc kubenswrapper[4935]: I1005 07:11:40.521910 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gjjk\" (UniqueName: \"kubernetes.io/projected/e98d30f9-d435-4c12-80c9-0f03a3853a0f-kube-api-access-5gjjk\") pod \"keystone-5ce4-account-create-n7jk5\" (UID: \"e98d30f9-d435-4c12-80c9-0f03a3853a0f\") " pod="openstack/keystone-5ce4-account-create-n7jk5" Oct 05 07:11:40 crc kubenswrapper[4935]: I1005 07:11:40.623602 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gjjk\" (UniqueName: \"kubernetes.io/projected/e98d30f9-d435-4c12-80c9-0f03a3853a0f-kube-api-access-5gjjk\") pod \"keystone-5ce4-account-create-n7jk5\" (UID: \"e98d30f9-d435-4c12-80c9-0f03a3853a0f\") " pod="openstack/keystone-5ce4-account-create-n7jk5" Oct 05 07:11:40 crc kubenswrapper[4935]: I1005 07:11:40.642115 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gjjk\" (UniqueName: \"kubernetes.io/projected/e98d30f9-d435-4c12-80c9-0f03a3853a0f-kube-api-access-5gjjk\") pod \"keystone-5ce4-account-create-n7jk5\" (UID: \"e98d30f9-d435-4c12-80c9-0f03a3853a0f\") " pod="openstack/keystone-5ce4-account-create-n7jk5" Oct 05 07:11:40 crc kubenswrapper[4935]: I1005 07:11:40.729682 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ce4-account-create-n7jk5" Oct 05 07:11:49 crc kubenswrapper[4935]: E1005 07:11:49.896477 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:b2f6e417260d4ae6bf70c05bf2243a0b6cac555cd28080772546761015674191" Oct 05 07:11:49 crc kubenswrapper[4935]: E1005 07:11:49.899469 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:b2f6e417260d4ae6bf70c05bf2243a0b6cac555cd28080772546761015674191,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-265rb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-mbbgc_openstack(2da110a9-7934-4fc0-88dc-8025c46413cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:11:49 crc kubenswrapper[4935]: E1005 07:11:49.901202 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-mbbgc" podUID="2da110a9-7934-4fc0-88dc-8025c46413cb" Oct 05 07:11:50 crc kubenswrapper[4935]: I1005 07:11:50.412464 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-b2q9j"] Oct 05 07:11:50 crc kubenswrapper[4935]: W1005 07:11:50.418944 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bf18d65_b3e6_41ef_a263_bd8642d272ed.slice/crio-a199f0934df186582e1d92b5dddf6aded98b52f675dbce525af215eb9142d02b WatchSource:0}: Error finding container a199f0934df186582e1d92b5dddf6aded98b52f675dbce525af215eb9142d02b: Status 404 returned error can't find the container with id a199f0934df186582e1d92b5dddf6aded98b52f675dbce525af215eb9142d02b Oct 05 07:11:50 crc kubenswrapper[4935]: I1005 07:11:50.428082 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"f8b9ec64a993b5fc267da34859f562afb461e51a8c3449b8a4980e4dbec60aab"} Oct 05 07:11:50 crc kubenswrapper[4935]: I1005 07:11:50.428144 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"b90d88185b137f6f204daabf2f63d9c6ca4190f72a5e20e02ad439fe47b1d9e2"} Oct 05 07:11:50 crc kubenswrapper[4935]: E1005 07:11:50.436933 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:b2f6e417260d4ae6bf70c05bf2243a0b6cac555cd28080772546761015674191\\\"\"" pod="openstack/glance-db-sync-mbbgc" podUID="2da110a9-7934-4fc0-88dc-8025c46413cb" Oct 05 07:11:50 crc kubenswrapper[4935]: I1005 07:11:50.519538 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-kpphs"] Oct 05 07:11:50 crc kubenswrapper[4935]: I1005 07:11:50.601264 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5z299"] Oct 05 07:11:50 crc kubenswrapper[4935]: I1005 07:11:50.637166 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5ce4-account-create-n7jk5"] Oct 05 07:11:50 crc kubenswrapper[4935]: W1005 07:11:50.639913 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7a3c603_bb1a_4a34_80e3_4cae0ef2815b.slice/crio-b69bfd1a9d333f9c32a6a289eabb7917968e948c64acae16cbcef1dd08c44729 WatchSource:0}: Error finding container b69bfd1a9d333f9c32a6a289eabb7917968e948c64acae16cbcef1dd08c44729: Status 404 returned error can't find the container with id b69bfd1a9d333f9c32a6a289eabb7917968e948c64acae16cbcef1dd08c44729 Oct 05 07:11:50 crc kubenswrapper[4935]: I1005 07:11:50.647441 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.439794 4935 generic.go:334] "Generic (PLEG): container finished" podID="8a278b3a-993f-427f-9ab5-d189cbfb4a69" containerID="158202d939bff1d98cf1e844dd1d81256f5a1daac2a99088ace4f7eb6d20c664" exitCode=0 Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.439867 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kpphs" event={"ID":"8a278b3a-993f-427f-9ab5-d189cbfb4a69","Type":"ContainerDied","Data":"158202d939bff1d98cf1e844dd1d81256f5a1daac2a99088ace4f7eb6d20c664"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.439915 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kpphs" event={"ID":"8a278b3a-993f-427f-9ab5-d189cbfb4a69","Type":"ContainerStarted","Data":"0c8627c36a902022ab363d44a395f49ebcc61a1f5c7af6a2677f679fc694e9a4"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.441781 4935 generic.go:334] "Generic (PLEG): container finished" podID="c7a3c603-bb1a-4a34-80e3-4cae0ef2815b" containerID="b721fb230c38e8b61034175f3484540666a622c814025c89af60628e2f544e6d" exitCode=0 Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.441837 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5z299" event={"ID":"c7a3c603-bb1a-4a34-80e3-4cae0ef2815b","Type":"ContainerDied","Data":"b721fb230c38e8b61034175f3484540666a622c814025c89af60628e2f544e6d"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.441856 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5z299" event={"ID":"c7a3c603-bb1a-4a34-80e3-4cae0ef2815b","Type":"ContainerStarted","Data":"b69bfd1a9d333f9c32a6a289eabb7917968e948c64acae16cbcef1dd08c44729"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.443660 4935 generic.go:334] "Generic (PLEG): container finished" podID="e98d30f9-d435-4c12-80c9-0f03a3853a0f" containerID="28bbaf182e3a7272bba91cd40617c6d28d901e11a85bd3ebc75a3a8af669c7b2" exitCode=0 Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.443723 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ce4-account-create-n7jk5" event={"ID":"e98d30f9-d435-4c12-80c9-0f03a3853a0f","Type":"ContainerDied","Data":"28bbaf182e3a7272bba91cd40617c6d28d901e11a85bd3ebc75a3a8af669c7b2"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.443749 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ce4-account-create-n7jk5" event={"ID":"e98d30f9-d435-4c12-80c9-0f03a3853a0f","Type":"ContainerStarted","Data":"cb2d20c036cbf124defdf5ed60e84d4f4e693e3bfbce828535dfa1c976058d1e"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.456586 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerStarted","Data":"fbfc0f5f81ecb328f21528e6e7fd64b97e302cbe3748716791576b9c52740ca0"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.460410 4935 generic.go:334] "Generic (PLEG): container finished" podID="0bf18d65-b3e6-41ef-a263-bd8642d272ed" containerID="adfb600248e13d2f6b1320de3c5a3560bd39190a465558e7111ae839f318c675" exitCode=0 Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.460466 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b2q9j" event={"ID":"0bf18d65-b3e6-41ef-a263-bd8642d272ed","Type":"ContainerDied","Data":"adfb600248e13d2f6b1320de3c5a3560bd39190a465558e7111ae839f318c675"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.460548 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b2q9j" event={"ID":"0bf18d65-b3e6-41ef-a263-bd8642d272ed","Type":"ContainerStarted","Data":"a199f0934df186582e1d92b5dddf6aded98b52f675dbce525af215eb9142d02b"} Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.594330 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=32.919897168 podStartE2EDuration="39.594299604s" podCreationTimestamp="2025-10-05 07:11:12 +0000 UTC" firstStartedPulling="2025-10-05 07:11:30.432789423 +0000 UTC m=+1124.315415883" lastFinishedPulling="2025-10-05 07:11:37.107191859 +0000 UTC m=+1130.989818319" observedRunningTime="2025-10-05 07:11:51.587946095 +0000 UTC m=+1145.470572595" watchObservedRunningTime="2025-10-05 07:11:51.594299604 +0000 UTC m=+1145.476926124" Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.923480 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77594f5649-cw86l"] Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.925158 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.931065 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 05 07:11:51 crc kubenswrapper[4935]: I1005 07:11:51.942213 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-cw86l"] Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.014287 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-sb\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.014394 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-config\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.014422 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-nb\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.014491 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-svc\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.014512 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-swift-storage-0\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.014625 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt589\" (UniqueName: \"kubernetes.io/projected/333447cb-9bb5-4e0f-bbd0-c320cecc108d-kube-api-access-vt589\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.115653 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-config\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.115708 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-nb\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.115776 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-svc\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.115798 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-swift-storage-0\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.115836 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt589\" (UniqueName: \"kubernetes.io/projected/333447cb-9bb5-4e0f-bbd0-c320cecc108d-kube-api-access-vt589\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.115917 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-sb\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.116880 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-sb\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.117052 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-svc\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.117635 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-config\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.117649 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-swift-storage-0\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.118014 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-nb\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.157210 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt589\" (UniqueName: \"kubernetes.io/projected/333447cb-9bb5-4e0f-bbd0-c320cecc108d-kube-api-access-vt589\") pod \"dnsmasq-dns-77594f5649-cw86l\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.247268 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.726376 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-cw86l"] Oct 05 07:11:52 crc kubenswrapper[4935]: W1005 07:11:52.756204 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333447cb_9bb5_4e0f_bbd0_c320cecc108d.slice/crio-d7a3841b39e2daded300aa80e8ef8e775b40e6c82db9f04d5d0c94f98a09b708 WatchSource:0}: Error finding container d7a3841b39e2daded300aa80e8ef8e775b40e6c82db9f04d5d0c94f98a09b708: Status 404 returned error can't find the container with id d7a3841b39e2daded300aa80e8ef8e775b40e6c82db9f04d5d0c94f98a09b708 Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.766657 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5z299" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.827579 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4p9s\" (UniqueName: \"kubernetes.io/projected/c7a3c603-bb1a-4a34-80e3-4cae0ef2815b-kube-api-access-c4p9s\") pod \"c7a3c603-bb1a-4a34-80e3-4cae0ef2815b\" (UID: \"c7a3c603-bb1a-4a34-80e3-4cae0ef2815b\") " Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.830722 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a3c603-bb1a-4a34-80e3-4cae0ef2815b-kube-api-access-c4p9s" (OuterVolumeSpecName: "kube-api-access-c4p9s") pod "c7a3c603-bb1a-4a34-80e3-4cae0ef2815b" (UID: "c7a3c603-bb1a-4a34-80e3-4cae0ef2815b"). InnerVolumeSpecName "kube-api-access-c4p9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.831439 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4p9s\" (UniqueName: \"kubernetes.io/projected/c7a3c603-bb1a-4a34-80e3-4cae0ef2815b-kube-api-access-c4p9s\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.851663 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b2q9j" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.857464 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kpphs" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.867270 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ce4-account-create-n7jk5" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.932263 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/8a278b3a-993f-427f-9ab5-d189cbfb4a69-kube-api-access-z2kcb\") pod \"8a278b3a-993f-427f-9ab5-d189cbfb4a69\" (UID: \"8a278b3a-993f-427f-9ab5-d189cbfb4a69\") " Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.932310 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gjjk\" (UniqueName: \"kubernetes.io/projected/e98d30f9-d435-4c12-80c9-0f03a3853a0f-kube-api-access-5gjjk\") pod \"e98d30f9-d435-4c12-80c9-0f03a3853a0f\" (UID: \"e98d30f9-d435-4c12-80c9-0f03a3853a0f\") " Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.932510 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dlnq\" (UniqueName: \"kubernetes.io/projected/0bf18d65-b3e6-41ef-a263-bd8642d272ed-kube-api-access-4dlnq\") pod \"0bf18d65-b3e6-41ef-a263-bd8642d272ed\" (UID: \"0bf18d65-b3e6-41ef-a263-bd8642d272ed\") " Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.935962 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf18d65-b3e6-41ef-a263-bd8642d272ed-kube-api-access-4dlnq" (OuterVolumeSpecName: "kube-api-access-4dlnq") pod "0bf18d65-b3e6-41ef-a263-bd8642d272ed" (UID: "0bf18d65-b3e6-41ef-a263-bd8642d272ed"). InnerVolumeSpecName "kube-api-access-4dlnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.936261 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98d30f9-d435-4c12-80c9-0f03a3853a0f-kube-api-access-5gjjk" (OuterVolumeSpecName: "kube-api-access-5gjjk") pod "e98d30f9-d435-4c12-80c9-0f03a3853a0f" (UID: "e98d30f9-d435-4c12-80c9-0f03a3853a0f"). InnerVolumeSpecName "kube-api-access-5gjjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:52 crc kubenswrapper[4935]: I1005 07:11:52.938820 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a278b3a-993f-427f-9ab5-d189cbfb4a69-kube-api-access-z2kcb" (OuterVolumeSpecName: "kube-api-access-z2kcb") pod "8a278b3a-993f-427f-9ab5-d189cbfb4a69" (UID: "8a278b3a-993f-427f-9ab5-d189cbfb4a69"). InnerVolumeSpecName "kube-api-access-z2kcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.034194 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dlnq\" (UniqueName: \"kubernetes.io/projected/0bf18d65-b3e6-41ef-a263-bd8642d272ed-kube-api-access-4dlnq\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.034229 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/8a278b3a-993f-427f-9ab5-d189cbfb4a69-kube-api-access-z2kcb\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.034238 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gjjk\" (UniqueName: \"kubernetes.io/projected/e98d30f9-d435-4c12-80c9-0f03a3853a0f-kube-api-access-5gjjk\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.483677 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ce4-account-create-n7jk5" event={"ID":"e98d30f9-d435-4c12-80c9-0f03a3853a0f","Type":"ContainerDied","Data":"cb2d20c036cbf124defdf5ed60e84d4f4e693e3bfbce828535dfa1c976058d1e"} Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.483741 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb2d20c036cbf124defdf5ed60e84d4f4e693e3bfbce828535dfa1c976058d1e" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.483743 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ce4-account-create-n7jk5" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.485384 4935 generic.go:334] "Generic (PLEG): container finished" podID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" containerID="c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38" exitCode=0 Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.485495 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-cw86l" event={"ID":"333447cb-9bb5-4e0f-bbd0-c320cecc108d","Type":"ContainerDied","Data":"c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38"} Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.485576 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-cw86l" event={"ID":"333447cb-9bb5-4e0f-bbd0-c320cecc108d","Type":"ContainerStarted","Data":"d7a3841b39e2daded300aa80e8ef8e775b40e6c82db9f04d5d0c94f98a09b708"} Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.490850 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b2q9j" event={"ID":"0bf18d65-b3e6-41ef-a263-bd8642d272ed","Type":"ContainerDied","Data":"a199f0934df186582e1d92b5dddf6aded98b52f675dbce525af215eb9142d02b"} Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.490876 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a199f0934df186582e1d92b5dddf6aded98b52f675dbce525af215eb9142d02b" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.490923 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b2q9j" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.494210 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kpphs" event={"ID":"8a278b3a-993f-427f-9ab5-d189cbfb4a69","Type":"ContainerDied","Data":"0c8627c36a902022ab363d44a395f49ebcc61a1f5c7af6a2677f679fc694e9a4"} Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.494272 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kpphs" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.494277 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c8627c36a902022ab363d44a395f49ebcc61a1f5c7af6a2677f679fc694e9a4" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.497073 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5z299" event={"ID":"c7a3c603-bb1a-4a34-80e3-4cae0ef2815b","Type":"ContainerDied","Data":"b69bfd1a9d333f9c32a6a289eabb7917968e948c64acae16cbcef1dd08c44729"} Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.497127 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b69bfd1a9d333f9c32a6a289eabb7917968e948c64acae16cbcef1dd08c44729" Oct 05 07:11:53 crc kubenswrapper[4935]: I1005 07:11:53.497203 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5z299" Oct 05 07:11:54 crc kubenswrapper[4935]: I1005 07:11:54.511396 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-cw86l" event={"ID":"333447cb-9bb5-4e0f-bbd0-c320cecc108d","Type":"ContainerStarted","Data":"4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e"} Oct 05 07:11:54 crc kubenswrapper[4935]: I1005 07:11:54.511988 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:11:54 crc kubenswrapper[4935]: I1005 07:11:54.549098 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77594f5649-cw86l" podStartSLOduration=3.549068642 podStartE2EDuration="3.549068642s" podCreationTimestamp="2025-10-05 07:11:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:11:54.537868575 +0000 UTC m=+1148.420495065" watchObservedRunningTime="2025-10-05 07:11:54.549068642 +0000 UTC m=+1148.431695132" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.020218 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-7s8ss"] Oct 05 07:11:56 crc kubenswrapper[4935]: E1005 07:11:56.020986 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a278b3a-993f-427f-9ab5-d189cbfb4a69" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.021010 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a278b3a-993f-427f-9ab5-d189cbfb4a69" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: E1005 07:11:56.021024 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a3c603-bb1a-4a34-80e3-4cae0ef2815b" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.021035 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a3c603-bb1a-4a34-80e3-4cae0ef2815b" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: E1005 07:11:56.021071 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf18d65-b3e6-41ef-a263-bd8642d272ed" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.021083 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf18d65-b3e6-41ef-a263-bd8642d272ed" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: E1005 07:11:56.021096 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98d30f9-d435-4c12-80c9-0f03a3853a0f" containerName="mariadb-account-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.021104 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98d30f9-d435-4c12-80c9-0f03a3853a0f" containerName="mariadb-account-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.021329 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a278b3a-993f-427f-9ab5-d189cbfb4a69" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.021352 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7a3c603-bb1a-4a34-80e3-4cae0ef2815b" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.021375 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98d30f9-d435-4c12-80c9-0f03a3853a0f" containerName="mariadb-account-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.021396 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf18d65-b3e6-41ef-a263-bd8642d272ed" containerName="mariadb-database-create" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.022128 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.026711 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xm842" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.026863 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.027289 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.027706 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.047344 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7s8ss"] Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.085409 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-combined-ca-bundle\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.085458 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snlhj\" (UniqueName: \"kubernetes.io/projected/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-kube-api-access-snlhj\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.085529 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-config-data\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.186930 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-config-data\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.187059 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-combined-ca-bundle\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.187097 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snlhj\" (UniqueName: \"kubernetes.io/projected/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-kube-api-access-snlhj\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.195907 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-combined-ca-bundle\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.195999 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-config-data\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.208123 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snlhj\" (UniqueName: \"kubernetes.io/projected/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-kube-api-access-snlhj\") pod \"keystone-db-sync-7s8ss\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:56 crc kubenswrapper[4935]: I1005 07:11:56.351073 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:11:57 crc kubenswrapper[4935]: I1005 07:11:57.472985 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7s8ss"] Oct 05 07:11:57 crc kubenswrapper[4935]: W1005 07:11:57.474266 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40c9b9aa_b26e_4b96_87a1_c7c82a1bffd8.slice/crio-34ed69f0908c4a9365761f18379896eec19307502feeae29136e608870293a72 WatchSource:0}: Error finding container 34ed69f0908c4a9365761f18379896eec19307502feeae29136e608870293a72: Status 404 returned error can't find the container with id 34ed69f0908c4a9365761f18379896eec19307502feeae29136e608870293a72 Oct 05 07:11:57 crc kubenswrapper[4935]: I1005 07:11:57.539282 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7s8ss" event={"ID":"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8","Type":"ContainerStarted","Data":"34ed69f0908c4a9365761f18379896eec19307502feeae29136e608870293a72"} Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.028249 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b178-account-create-tl96t"] Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.030835 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b178-account-create-tl96t" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.033130 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.041362 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b178-account-create-tl96t"] Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.145701 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6nxj\" (UniqueName: \"kubernetes.io/projected/5bff7991-7fc7-4034-82dd-bc71fc7b3f7a-kube-api-access-v6nxj\") pod \"barbican-b178-account-create-tl96t\" (UID: \"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a\") " pod="openstack/barbican-b178-account-create-tl96t" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.233654 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9bb0-account-create-g9ffr"] Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.240397 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9bb0-account-create-g9ffr" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.242248 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9bb0-account-create-g9ffr"] Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.244790 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.249031 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6nxj\" (UniqueName: \"kubernetes.io/projected/5bff7991-7fc7-4034-82dd-bc71fc7b3f7a-kube-api-access-v6nxj\") pod \"barbican-b178-account-create-tl96t\" (UID: \"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a\") " pod="openstack/barbican-b178-account-create-tl96t" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.273380 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6nxj\" (UniqueName: \"kubernetes.io/projected/5bff7991-7fc7-4034-82dd-bc71fc7b3f7a-kube-api-access-v6nxj\") pod \"barbican-b178-account-create-tl96t\" (UID: \"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a\") " pod="openstack/barbican-b178-account-create-tl96t" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.333395 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b3bb-account-create-pddl8"] Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.338163 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b3bb-account-create-pddl8" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.343938 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.352187 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b3bb-account-create-pddl8"] Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.354146 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66sx9\" (UniqueName: \"kubernetes.io/projected/1d21608a-d4fa-4bba-924f-2f62ff4e8ebd-kube-api-access-66sx9\") pod \"cinder-9bb0-account-create-g9ffr\" (UID: \"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd\") " pod="openstack/cinder-9bb0-account-create-g9ffr" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.363330 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b178-account-create-tl96t" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.456506 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66sx9\" (UniqueName: \"kubernetes.io/projected/1d21608a-d4fa-4bba-924f-2f62ff4e8ebd-kube-api-access-66sx9\") pod \"cinder-9bb0-account-create-g9ffr\" (UID: \"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd\") " pod="openstack/cinder-9bb0-account-create-g9ffr" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.456569 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtztk\" (UniqueName: \"kubernetes.io/projected/115ec2f7-7585-4a18-916b-7e253535e497-kube-api-access-gtztk\") pod \"neutron-b3bb-account-create-pddl8\" (UID: \"115ec2f7-7585-4a18-916b-7e253535e497\") " pod="openstack/neutron-b3bb-account-create-pddl8" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.475572 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66sx9\" (UniqueName: \"kubernetes.io/projected/1d21608a-d4fa-4bba-924f-2f62ff4e8ebd-kube-api-access-66sx9\") pod \"cinder-9bb0-account-create-g9ffr\" (UID: \"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd\") " pod="openstack/cinder-9bb0-account-create-g9ffr" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.557725 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtztk\" (UniqueName: \"kubernetes.io/projected/115ec2f7-7585-4a18-916b-7e253535e497-kube-api-access-gtztk\") pod \"neutron-b3bb-account-create-pddl8\" (UID: \"115ec2f7-7585-4a18-916b-7e253535e497\") " pod="openstack/neutron-b3bb-account-create-pddl8" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.566424 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9bb0-account-create-g9ffr" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.578661 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtztk\" (UniqueName: \"kubernetes.io/projected/115ec2f7-7585-4a18-916b-7e253535e497-kube-api-access-gtztk\") pod \"neutron-b3bb-account-create-pddl8\" (UID: \"115ec2f7-7585-4a18-916b-7e253535e497\") " pod="openstack/neutron-b3bb-account-create-pddl8" Oct 05 07:11:59 crc kubenswrapper[4935]: I1005 07:11:59.663229 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b3bb-account-create-pddl8" Oct 05 07:12:01 crc kubenswrapper[4935]: I1005 07:12:01.935431 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b3bb-account-create-pddl8"] Oct 05 07:12:01 crc kubenswrapper[4935]: W1005 07:12:01.937306 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod115ec2f7_7585_4a18_916b_7e253535e497.slice/crio-c6bf75fcd78f7cd21251c33daf2ad94224c19901e3219218845e745d4ce40bd0 WatchSource:0}: Error finding container c6bf75fcd78f7cd21251c33daf2ad94224c19901e3219218845e745d4ce40bd0: Status 404 returned error can't find the container with id c6bf75fcd78f7cd21251c33daf2ad94224c19901e3219218845e745d4ce40bd0 Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.006304 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b178-account-create-tl96t"] Oct 05 07:12:02 crc kubenswrapper[4935]: W1005 07:12:02.015465 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bff7991_7fc7_4034_82dd_bc71fc7b3f7a.slice/crio-05c48ad06a457dd2da3abb0694d0a9884b2aa438489e1f2f7ffdd829d1d40e73 WatchSource:0}: Error finding container 05c48ad06a457dd2da3abb0694d0a9884b2aa438489e1f2f7ffdd829d1d40e73: Status 404 returned error can't find the container with id 05c48ad06a457dd2da3abb0694d0a9884b2aa438489e1f2f7ffdd829d1d40e73 Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.071438 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9bb0-account-create-g9ffr"] Oct 05 07:12:02 crc kubenswrapper[4935]: W1005 07:12:02.072638 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d21608a_d4fa_4bba_924f_2f62ff4e8ebd.slice/crio-ed3883a94f43c1c9373bffafad1b5fe1785b3e649343ad7a94412e605dd11043 WatchSource:0}: Error finding container ed3883a94f43c1c9373bffafad1b5fe1785b3e649343ad7a94412e605dd11043: Status 404 returned error can't find the container with id ed3883a94f43c1c9373bffafad1b5fe1785b3e649343ad7a94412e605dd11043 Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.250346 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.309134 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-vvdbb"] Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.309763 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" podUID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerName="dnsmasq-dns" containerID="cri-o://cdae626b037a8bbd125e95275e0ed740f42fd9e0890150e0768a66703b004ff3" gracePeriod=10 Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.592874 4935 generic.go:334] "Generic (PLEG): container finished" podID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerID="cdae626b037a8bbd125e95275e0ed740f42fd9e0890150e0768a66703b004ff3" exitCode=0 Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.592929 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" event={"ID":"116b1bb7-6886-431a-9127-41e40a8b1fd4","Type":"ContainerDied","Data":"cdae626b037a8bbd125e95275e0ed740f42fd9e0890150e0768a66703b004ff3"} Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.594713 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9bb0-account-create-g9ffr" event={"ID":"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd","Type":"ContainerStarted","Data":"7cf0b54a3199310d1b75ab3b5890bce9c54028af4fb2c5cde26e65b39f123076"} Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.594781 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9bb0-account-create-g9ffr" event={"ID":"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd","Type":"ContainerStarted","Data":"ed3883a94f43c1c9373bffafad1b5fe1785b3e649343ad7a94412e605dd11043"} Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.596218 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b3bb-account-create-pddl8" event={"ID":"115ec2f7-7585-4a18-916b-7e253535e497","Type":"ContainerStarted","Data":"4b10ad46cc6f92c33279323585a2011416f051915eeaaf3790fc0bd4cbf3080f"} Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.596257 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b3bb-account-create-pddl8" event={"ID":"115ec2f7-7585-4a18-916b-7e253535e497","Type":"ContainerStarted","Data":"c6bf75fcd78f7cd21251c33daf2ad94224c19901e3219218845e745d4ce40bd0"} Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.598413 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7s8ss" event={"ID":"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8","Type":"ContainerStarted","Data":"221fe2b717870808d7d0e598673cb75e96ed270806b646d7a5d1848b41337223"} Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.601599 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b178-account-create-tl96t" event={"ID":"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a","Type":"ContainerStarted","Data":"5921e4650b95f588079eb80fa686b2459174b8c9e94cdf90d7ec54954b6b541d"} Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.601655 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b178-account-create-tl96t" event={"ID":"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a","Type":"ContainerStarted","Data":"05c48ad06a457dd2da3abb0694d0a9884b2aa438489e1f2f7ffdd829d1d40e73"} Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.613266 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-9bb0-account-create-g9ffr" podStartSLOduration=3.6132444919999998 podStartE2EDuration="3.613244492s" podCreationTimestamp="2025-10-05 07:11:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:02.611171917 +0000 UTC m=+1156.493798377" watchObservedRunningTime="2025-10-05 07:12:02.613244492 +0000 UTC m=+1156.495870952" Oct 05 07:12:02 crc kubenswrapper[4935]: I1005 07:12:02.626853 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-7s8ss" podStartSLOduration=3.604594398 podStartE2EDuration="7.626833602s" podCreationTimestamp="2025-10-05 07:11:55 +0000 UTC" firstStartedPulling="2025-10-05 07:11:57.480026419 +0000 UTC m=+1151.362652879" lastFinishedPulling="2025-10-05 07:12:01.502265613 +0000 UTC m=+1155.384892083" observedRunningTime="2025-10-05 07:12:02.623330059 +0000 UTC m=+1156.505956519" watchObservedRunningTime="2025-10-05 07:12:02.626833602 +0000 UTC m=+1156.509460062" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.029348 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" podUID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.545223 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.613404 4935 generic.go:334] "Generic (PLEG): container finished" podID="1d21608a-d4fa-4bba-924f-2f62ff4e8ebd" containerID="7cf0b54a3199310d1b75ab3b5890bce9c54028af4fb2c5cde26e65b39f123076" exitCode=0 Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.613825 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9bb0-account-create-g9ffr" event={"ID":"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd","Type":"ContainerDied","Data":"7cf0b54a3199310d1b75ab3b5890bce9c54028af4fb2c5cde26e65b39f123076"} Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.616730 4935 generic.go:334] "Generic (PLEG): container finished" podID="115ec2f7-7585-4a18-916b-7e253535e497" containerID="4b10ad46cc6f92c33279323585a2011416f051915eeaaf3790fc0bd4cbf3080f" exitCode=0 Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.616818 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b3bb-account-create-pddl8" event={"ID":"115ec2f7-7585-4a18-916b-7e253535e497","Type":"ContainerDied","Data":"4b10ad46cc6f92c33279323585a2011416f051915eeaaf3790fc0bd4cbf3080f"} Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.619983 4935 generic.go:334] "Generic (PLEG): container finished" podID="5bff7991-7fc7-4034-82dd-bc71fc7b3f7a" containerID="5921e4650b95f588079eb80fa686b2459174b8c9e94cdf90d7ec54954b6b541d" exitCode=0 Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.620041 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b178-account-create-tl96t" event={"ID":"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a","Type":"ContainerDied","Data":"5921e4650b95f588079eb80fa686b2459174b8c9e94cdf90d7ec54954b6b541d"} Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.625939 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.626010 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-vvdbb" event={"ID":"116b1bb7-6886-431a-9127-41e40a8b1fd4","Type":"ContainerDied","Data":"cdc7af747c4ea3c69ccaf5c3d22e1dacb5833052866ebc1a58b30e2d77181de5"} Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.626134 4935 scope.go:117] "RemoveContainer" containerID="cdae626b037a8bbd125e95275e0ed740f42fd9e0890150e0768a66703b004ff3" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.662517 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-sb\") pod \"116b1bb7-6886-431a-9127-41e40a8b1fd4\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.663270 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvqmz\" (UniqueName: \"kubernetes.io/projected/116b1bb7-6886-431a-9127-41e40a8b1fd4-kube-api-access-pvqmz\") pod \"116b1bb7-6886-431a-9127-41e40a8b1fd4\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.663302 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-dns-svc\") pod \"116b1bb7-6886-431a-9127-41e40a8b1fd4\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.663362 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-config\") pod \"116b1bb7-6886-431a-9127-41e40a8b1fd4\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.663413 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-nb\") pod \"116b1bb7-6886-431a-9127-41e40a8b1fd4\" (UID: \"116b1bb7-6886-431a-9127-41e40a8b1fd4\") " Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.667774 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/116b1bb7-6886-431a-9127-41e40a8b1fd4-kube-api-access-pvqmz" (OuterVolumeSpecName: "kube-api-access-pvqmz") pod "116b1bb7-6886-431a-9127-41e40a8b1fd4" (UID: "116b1bb7-6886-431a-9127-41e40a8b1fd4"). InnerVolumeSpecName "kube-api-access-pvqmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.708688 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "116b1bb7-6886-431a-9127-41e40a8b1fd4" (UID: "116b1bb7-6886-431a-9127-41e40a8b1fd4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.712447 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-config" (OuterVolumeSpecName: "config") pod "116b1bb7-6886-431a-9127-41e40a8b1fd4" (UID: "116b1bb7-6886-431a-9127-41e40a8b1fd4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.725415 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "116b1bb7-6886-431a-9127-41e40a8b1fd4" (UID: "116b1bb7-6886-431a-9127-41e40a8b1fd4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.725626 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "116b1bb7-6886-431a-9127-41e40a8b1fd4" (UID: "116b1bb7-6886-431a-9127-41e40a8b1fd4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.765592 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvqmz\" (UniqueName: \"kubernetes.io/projected/116b1bb7-6886-431a-9127-41e40a8b1fd4-kube-api-access-pvqmz\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.765629 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.765641 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.765650 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.765659 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/116b1bb7-6886-431a-9127-41e40a8b1fd4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.810573 4935 scope.go:117] "RemoveContainer" containerID="3e7174750dde9789e06e2dd8200779016b4bc5330ebc8df3b13f82263e262821" Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.963846 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-vvdbb"] Oct 05 07:12:03 crc kubenswrapper[4935]: I1005 07:12:03.971694 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-vvdbb"] Oct 05 07:12:04 crc kubenswrapper[4935]: I1005 07:12:04.643160 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mbbgc" event={"ID":"2da110a9-7934-4fc0-88dc-8025c46413cb","Type":"ContainerStarted","Data":"e61a98d7b38be24992c47d287153ee58d5df1cadd58c8589fb3c0f806844634c"} Oct 05 07:12:04 crc kubenswrapper[4935]: I1005 07:12:04.786808 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="116b1bb7-6886-431a-9127-41e40a8b1fd4" path="/var/lib/kubelet/pods/116b1bb7-6886-431a-9127-41e40a8b1fd4/volumes" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.056704 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9bb0-account-create-g9ffr" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.061607 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b178-account-create-tl96t" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.065677 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b3bb-account-create-pddl8" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.079689 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-mbbgc" podStartSLOduration=2.484053092 podStartE2EDuration="34.079662057s" podCreationTimestamp="2025-10-05 07:11:31 +0000 UTC" firstStartedPulling="2025-10-05 07:11:31.94063093 +0000 UTC m=+1125.823257390" lastFinishedPulling="2025-10-05 07:12:03.536239895 +0000 UTC m=+1157.418866355" observedRunningTime="2025-10-05 07:12:04.670675818 +0000 UTC m=+1158.553302278" watchObservedRunningTime="2025-10-05 07:12:05.079662057 +0000 UTC m=+1158.962288517" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.099137 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66sx9\" (UniqueName: \"kubernetes.io/projected/1d21608a-d4fa-4bba-924f-2f62ff4e8ebd-kube-api-access-66sx9\") pod \"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd\" (UID: \"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd\") " Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.107813 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d21608a-d4fa-4bba-924f-2f62ff4e8ebd-kube-api-access-66sx9" (OuterVolumeSpecName: "kube-api-access-66sx9") pod "1d21608a-d4fa-4bba-924f-2f62ff4e8ebd" (UID: "1d21608a-d4fa-4bba-924f-2f62ff4e8ebd"). InnerVolumeSpecName "kube-api-access-66sx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.200268 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtztk\" (UniqueName: \"kubernetes.io/projected/115ec2f7-7585-4a18-916b-7e253535e497-kube-api-access-gtztk\") pod \"115ec2f7-7585-4a18-916b-7e253535e497\" (UID: \"115ec2f7-7585-4a18-916b-7e253535e497\") " Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.200544 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6nxj\" (UniqueName: \"kubernetes.io/projected/5bff7991-7fc7-4034-82dd-bc71fc7b3f7a-kube-api-access-v6nxj\") pod \"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a\" (UID: \"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a\") " Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.202120 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66sx9\" (UniqueName: \"kubernetes.io/projected/1d21608a-d4fa-4bba-924f-2f62ff4e8ebd-kube-api-access-66sx9\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.204028 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/115ec2f7-7585-4a18-916b-7e253535e497-kube-api-access-gtztk" (OuterVolumeSpecName: "kube-api-access-gtztk") pod "115ec2f7-7585-4a18-916b-7e253535e497" (UID: "115ec2f7-7585-4a18-916b-7e253535e497"). InnerVolumeSpecName "kube-api-access-gtztk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.204171 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bff7991-7fc7-4034-82dd-bc71fc7b3f7a-kube-api-access-v6nxj" (OuterVolumeSpecName: "kube-api-access-v6nxj") pod "5bff7991-7fc7-4034-82dd-bc71fc7b3f7a" (UID: "5bff7991-7fc7-4034-82dd-bc71fc7b3f7a"). InnerVolumeSpecName "kube-api-access-v6nxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.303684 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6nxj\" (UniqueName: \"kubernetes.io/projected/5bff7991-7fc7-4034-82dd-bc71fc7b3f7a-kube-api-access-v6nxj\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.304144 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtztk\" (UniqueName: \"kubernetes.io/projected/115ec2f7-7585-4a18-916b-7e253535e497-kube-api-access-gtztk\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.653519 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b3bb-account-create-pddl8" event={"ID":"115ec2f7-7585-4a18-916b-7e253535e497","Type":"ContainerDied","Data":"c6bf75fcd78f7cd21251c33daf2ad94224c19901e3219218845e745d4ce40bd0"} Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.653569 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6bf75fcd78f7cd21251c33daf2ad94224c19901e3219218845e745d4ce40bd0" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.654704 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b3bb-account-create-pddl8" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.656511 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b178-account-create-tl96t" event={"ID":"5bff7991-7fc7-4034-82dd-bc71fc7b3f7a","Type":"ContainerDied","Data":"05c48ad06a457dd2da3abb0694d0a9884b2aa438489e1f2f7ffdd829d1d40e73"} Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.656541 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05c48ad06a457dd2da3abb0694d0a9884b2aa438489e1f2f7ffdd829d1d40e73" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.656543 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b178-account-create-tl96t" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.658506 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9bb0-account-create-g9ffr" event={"ID":"1d21608a-d4fa-4bba-924f-2f62ff4e8ebd","Type":"ContainerDied","Data":"ed3883a94f43c1c9373bffafad1b5fe1785b3e649343ad7a94412e605dd11043"} Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.658558 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed3883a94f43c1c9373bffafad1b5fe1785b3e649343ad7a94412e605dd11043" Oct 05 07:12:05 crc kubenswrapper[4935]: I1005 07:12:05.658562 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9bb0-account-create-g9ffr" Oct 05 07:12:06 crc kubenswrapper[4935]: I1005 07:12:06.667844 4935 generic.go:334] "Generic (PLEG): container finished" podID="40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8" containerID="221fe2b717870808d7d0e598673cb75e96ed270806b646d7a5d1848b41337223" exitCode=0 Oct 05 07:12:06 crc kubenswrapper[4935]: I1005 07:12:06.667930 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7s8ss" event={"ID":"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8","Type":"ContainerDied","Data":"221fe2b717870808d7d0e598673cb75e96ed270806b646d7a5d1848b41337223"} Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:07.999518 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.048280 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-config-data\") pod \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.048692 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-combined-ca-bundle\") pod \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.048830 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snlhj\" (UniqueName: \"kubernetes.io/projected/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-kube-api-access-snlhj\") pod \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\" (UID: \"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8\") " Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.054111 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-kube-api-access-snlhj" (OuterVolumeSpecName: "kube-api-access-snlhj") pod "40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8" (UID: "40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8"). InnerVolumeSpecName "kube-api-access-snlhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.132028 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8" (UID: "40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.146532 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-config-data" (OuterVolumeSpecName: "config-data") pod "40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8" (UID: "40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.151081 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.151103 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.151127 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snlhj\" (UniqueName: \"kubernetes.io/projected/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8-kube-api-access-snlhj\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.689836 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7s8ss" event={"ID":"40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8","Type":"ContainerDied","Data":"34ed69f0908c4a9365761f18379896eec19307502feeae29136e608870293a72"} Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.689942 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34ed69f0908c4a9365761f18379896eec19307502feeae29136e608870293a72" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.689957 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7s8ss" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.985616 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd8c5876c-qxqvb"] Oct 05 07:12:08 crc kubenswrapper[4935]: E1005 07:12:08.986055 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8" containerName="keystone-db-sync" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986068 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8" containerName="keystone-db-sync" Oct 05 07:12:08 crc kubenswrapper[4935]: E1005 07:12:08.986078 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d21608a-d4fa-4bba-924f-2f62ff4e8ebd" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986084 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d21608a-d4fa-4bba-924f-2f62ff4e8ebd" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: E1005 07:12:08.986099 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerName="init" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986106 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerName="init" Oct 05 07:12:08 crc kubenswrapper[4935]: E1005 07:12:08.986119 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerName="dnsmasq-dns" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986125 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerName="dnsmasq-dns" Oct 05 07:12:08 crc kubenswrapper[4935]: E1005 07:12:08.986142 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bff7991-7fc7-4034-82dd-bc71fc7b3f7a" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986147 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bff7991-7fc7-4034-82dd-bc71fc7b3f7a" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: E1005 07:12:08.986159 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115ec2f7-7585-4a18-916b-7e253535e497" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986164 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="115ec2f7-7585-4a18-916b-7e253535e497" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986345 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d21608a-d4fa-4bba-924f-2f62ff4e8ebd" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986360 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bff7991-7fc7-4034-82dd-bc71fc7b3f7a" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986378 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8" containerName="keystone-db-sync" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986386 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="116b1bb7-6886-431a-9127-41e40a8b1fd4" containerName="dnsmasq-dns" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.986394 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="115ec2f7-7585-4a18-916b-7e253535e497" containerName="mariadb-account-create" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.987259 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:08 crc kubenswrapper[4935]: I1005 07:12:08.995633 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5pgwt"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.002738 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.008692 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.008939 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.009084 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xm842" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.009262 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd8c5876c-qxqvb"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.013355 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.025641 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5pgwt"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.068925 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-config\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069214 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k8q8\" (UniqueName: \"kubernetes.io/projected/188231fb-5966-4618-8942-270277384480-kube-api-access-7k8q8\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069249 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069269 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq7x5\" (UniqueName: \"kubernetes.io/projected/ff3243ad-1c09-4957-8eb6-2413c8132dd4-kube-api-access-xq7x5\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069288 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-combined-ca-bundle\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069327 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-config-data\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069353 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-svc\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069376 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-credential-keys\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069398 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-fernet-keys\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069422 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-scripts\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069437 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.069562 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.170942 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-config\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.170993 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k8q8\" (UniqueName: \"kubernetes.io/projected/188231fb-5966-4618-8942-270277384480-kube-api-access-7k8q8\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171030 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171051 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq7x5\" (UniqueName: \"kubernetes.io/projected/ff3243ad-1c09-4957-8eb6-2413c8132dd4-kube-api-access-xq7x5\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171067 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-combined-ca-bundle\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171099 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-config-data\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171126 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-svc\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171148 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-credential-keys\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171168 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-fernet-keys\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171196 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-scripts\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171212 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.171248 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.172045 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.172530 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-config\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.173270 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.174814 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.174814 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-svc\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.180878 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-config-data\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.182656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-fernet-keys\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.189269 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-scripts\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.209769 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-credential-keys\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.210140 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-combined-ca-bundle\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.213588 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k8q8\" (UniqueName: \"kubernetes.io/projected/188231fb-5966-4618-8942-270277384480-kube-api-access-7k8q8\") pod \"keystone-bootstrap-5pgwt\" (UID: \"188231fb-5966-4618-8942-270277384480\") " pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.236980 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.241053 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.241697 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq7x5\" (UniqueName: \"kubernetes.io/projected/ff3243ad-1c09-4957-8eb6-2413c8132dd4-kube-api-access-xq7x5\") pod \"dnsmasq-dns-5cd8c5876c-qxqvb\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.253338 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd8c5876c-qxqvb"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.253928 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.277004 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.277249 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.304935 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-2vfdz"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.317266 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.322084 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.325473 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vbqqd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.325690 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.325797 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.336256 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.355447 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2vfdz"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.370328 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dcc79c769-qshg5"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.371753 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378593 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsjtt\" (UniqueName: \"kubernetes.io/projected/6e808378-4679-42ed-a47c-033e1edfb362-kube-api-access-bsjtt\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378642 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-config-data\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378666 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378680 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-scripts\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378723 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-scripts\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378737 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2f2s\" (UniqueName: \"kubernetes.io/projected/8c71f889-f8cd-4f42-a125-faaaaa1d3501-kube-api-access-s2f2s\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378763 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-config-data\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378780 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e808378-4679-42ed-a47c-033e1edfb362-logs\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378798 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-run-httpd\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378828 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-combined-ca-bundle\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378864 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-log-httpd\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.378913 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.400949 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dcc79c769-qshg5"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.469554 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-n9vr5"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.470860 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.483880 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.483991 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-scripts\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484025 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2f2s\" (UniqueName: \"kubernetes.io/projected/8c71f889-f8cd-4f42-a125-faaaaa1d3501-kube-api-access-s2f2s\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484060 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-config-data\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484076 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-tpbjr" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484178 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484463 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e808378-4679-42ed-a47c-033e1edfb362-logs\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484076 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e808378-4679-42ed-a47c-033e1edfb362-logs\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484679 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-run-httpd\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484760 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-config\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484794 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-combined-ca-bundle\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484843 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdxc7\" (UniqueName: \"kubernetes.io/projected/30631f10-b2bd-4f3a-9899-54f255c18222-kube-api-access-xdxc7\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484876 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-svc\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.484984 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-sb\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.485021 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-log-httpd\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.485162 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-swift-storage-0\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.485190 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-nb\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.490832 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-config-data\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.491570 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-log-httpd\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.491955 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-run-httpd\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.492396 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-scripts\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.493013 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.493511 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsjtt\" (UniqueName: \"kubernetes.io/projected/6e808378-4679-42ed-a47c-033e1edfb362-kube-api-access-bsjtt\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.493535 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-config-data\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.493583 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-scripts\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.493604 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.501715 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-n9vr5"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.503565 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-scripts\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.511360 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-combined-ca-bundle\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.520028 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.520089 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-config-data\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.520324 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.527687 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsjtt\" (UniqueName: \"kubernetes.io/projected/6e808378-4679-42ed-a47c-033e1edfb362-kube-api-access-bsjtt\") pod \"placement-db-sync-2vfdz\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.532469 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-nl9vd"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.536070 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.536431 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2f2s\" (UniqueName: \"kubernetes.io/projected/8c71f889-f8cd-4f42-a125-faaaaa1d3501-kube-api-access-s2f2s\") pod \"ceilometer-0\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.539009 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.539173 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x7t8h" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.546419 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nl9vd"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.586181 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-5fwl7"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.588838 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.591393 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.591940 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8g6z4" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.598760 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.601677 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-db-sync-config-data\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.601754 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-combined-ca-bundle\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.601799 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-scripts\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.601829 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-combined-ca-bundle\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.605111 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-config\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.605197 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skwwn\" (UniqueName: \"kubernetes.io/projected/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-kube-api-access-skwwn\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.605335 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdxc7\" (UniqueName: \"kubernetes.io/projected/30631f10-b2bd-4f3a-9899-54f255c18222-kube-api-access-xdxc7\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.605374 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z26l2\" (UniqueName: \"kubernetes.io/projected/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-kube-api-access-z26l2\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.605414 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-svc\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.606390 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-config\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.606559 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-svc\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.606862 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-etc-machine-id\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.606929 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-sb\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.607076 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-swift-storage-0\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.607117 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-nb\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.607875 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-swift-storage-0\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.608185 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-config-data\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.608357 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-db-sync-config-data\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.608449 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-sb\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.608840 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-nb\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.625702 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5fwl7"] Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.630212 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdxc7\" (UniqueName: \"kubernetes.io/projected/30631f10-b2bd-4f3a-9899-54f255c18222-kube-api-access-xdxc7\") pod \"dnsmasq-dns-dcc79c769-qshg5\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.698436 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711519 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z22w8\" (UniqueName: \"kubernetes.io/projected/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-kube-api-access-z22w8\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711634 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-config-data\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711661 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-db-sync-config-data\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711703 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-db-sync-config-data\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711725 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-combined-ca-bundle\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711743 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-scripts\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711781 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-combined-ca-bundle\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711801 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-config\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711821 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skwwn\" (UniqueName: \"kubernetes.io/projected/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-kube-api-access-skwwn\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711841 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-combined-ca-bundle\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711873 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z26l2\" (UniqueName: \"kubernetes.io/projected/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-kube-api-access-z26l2\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711915 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-etc-machine-id\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.711994 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-etc-machine-id\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.715955 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-config-data\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.717638 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-scripts\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.717805 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-combined-ca-bundle\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.719181 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-db-sync-config-data\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.721319 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-db-sync-config-data\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.721693 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.724221 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-combined-ca-bundle\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.730161 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skwwn\" (UniqueName: \"kubernetes.io/projected/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-kube-api-access-skwwn\") pod \"barbican-db-sync-nl9vd\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.731478 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z26l2\" (UniqueName: \"kubernetes.io/projected/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-kube-api-access-z26l2\") pod \"cinder-db-sync-n9vr5\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.747272 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.814298 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-config\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.814340 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-combined-ca-bundle\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.814385 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z22w8\" (UniqueName: \"kubernetes.io/projected/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-kube-api-access-z22w8\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.824838 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-config\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.833594 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-combined-ca-bundle\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.834016 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.848327 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z22w8\" (UniqueName: \"kubernetes.io/projected/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-kube-api-access-z22w8\") pod \"neutron-db-sync-5fwl7\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.866470 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.924420 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:09 crc kubenswrapper[4935]: I1005 07:12:09.934745 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd8c5876c-qxqvb"] Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.105751 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5pgwt"] Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.209427 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:10 crc kubenswrapper[4935]: W1005 07:12:10.237644 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c71f889_f8cd_4f42_a125_faaaaa1d3501.slice/crio-e47551e80dc05743699b4d5fa2006d5e7f6fd8a591227c880cc7668f115e2e0c WatchSource:0}: Error finding container e47551e80dc05743699b4d5fa2006d5e7f6fd8a591227c880cc7668f115e2e0c: Status 404 returned error can't find the container with id e47551e80dc05743699b4d5fa2006d5e7f6fd8a591227c880cc7668f115e2e0c Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.382645 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dcc79c769-qshg5"] Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.506656 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2vfdz"] Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.625851 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nl9vd"] Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.652578 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-n9vr5"] Oct 05 07:12:10 crc kubenswrapper[4935]: W1005 07:12:10.652909 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9ba8cbb_45b9_49e5_b5a0_9277b724b607.slice/crio-54666955fce9b36f2a70f3c5c88e9a86a09098ad74e7e05f6db9dcb24c298d53 WatchSource:0}: Error finding container 54666955fce9b36f2a70f3c5c88e9a86a09098ad74e7e05f6db9dcb24c298d53: Status 404 returned error can't find the container with id 54666955fce9b36f2a70f3c5c88e9a86a09098ad74e7e05f6db9dcb24c298d53 Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.723139 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nl9vd" event={"ID":"2ef780aa-5f7e-4fa2-a019-6ee2c3993692","Type":"ContainerStarted","Data":"7227578270f1512894036edc79f46ad87e6b2db92de39cacf28ea2b2be313fb5"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.725992 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5pgwt" event={"ID":"188231fb-5966-4618-8942-270277384480","Type":"ContainerStarted","Data":"4ac38cab9550b15f03bb7665210b5060057bc358f157b890ebd6d4e945ea3261"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.726022 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5pgwt" event={"ID":"188231fb-5966-4618-8942-270277384480","Type":"ContainerStarted","Data":"02309f5b1bb39825156a62a6640b0bf9ad8b9e57c7f78315ccc8c41bef2f9af0"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.734125 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" event={"ID":"30631f10-b2bd-4f3a-9899-54f255c18222","Type":"ContainerStarted","Data":"2d11f160ab6b209c7c31e74f226a72599200f3187ed78af7b2bfbe26c437ee1a"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.734166 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" event={"ID":"30631f10-b2bd-4f3a-9899-54f255c18222","Type":"ContainerStarted","Data":"cfcfeca7fe30a1f4f41a4453250d850c8fb34760875eda0e3e46a93cdd3e9446"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.735555 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vfdz" event={"ID":"6e808378-4679-42ed-a47c-033e1edfb362","Type":"ContainerStarted","Data":"096f3674640cbb85a5df9f4b6849b5a89a5b0ddfcda84e8ff239d0fb6d4f32d7"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.739059 4935 generic.go:334] "Generic (PLEG): container finished" podID="ff3243ad-1c09-4957-8eb6-2413c8132dd4" containerID="c186f5d7b8259988331ccedc07be2700f0d34e70cc3e5fb3de11ddc492095f7a" exitCode=0 Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.739114 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" event={"ID":"ff3243ad-1c09-4957-8eb6-2413c8132dd4","Type":"ContainerDied","Data":"c186f5d7b8259988331ccedc07be2700f0d34e70cc3e5fb3de11ddc492095f7a"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.739140 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" event={"ID":"ff3243ad-1c09-4957-8eb6-2413c8132dd4","Type":"ContainerStarted","Data":"a9c1491cea0a47f0c038e6ef3131e2004188f6e737001f40bcbd5837d4f1e5e0"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.748143 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5pgwt" podStartSLOduration=2.7481264579999998 podStartE2EDuration="2.748126458s" podCreationTimestamp="2025-10-05 07:12:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:10.746012982 +0000 UTC m=+1164.628639432" watchObservedRunningTime="2025-10-05 07:12:10.748126458 +0000 UTC m=+1164.630752918" Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.749118 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n9vr5" event={"ID":"e9ba8cbb-45b9-49e5-b5a0-9277b724b607","Type":"ContainerStarted","Data":"54666955fce9b36f2a70f3c5c88e9a86a09098ad74e7e05f6db9dcb24c298d53"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.750824 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerStarted","Data":"e47551e80dc05743699b4d5fa2006d5e7f6fd8a591227c880cc7668f115e2e0c"} Oct 05 07:12:10 crc kubenswrapper[4935]: I1005 07:12:10.818608 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5fwl7"] Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.074373 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.156058 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-sb\") pod \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.156132 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq7x5\" (UniqueName: \"kubernetes.io/projected/ff3243ad-1c09-4957-8eb6-2413c8132dd4-kube-api-access-xq7x5\") pod \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.156174 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-swift-storage-0\") pod \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.156194 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-nb\") pod \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.156225 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-config\") pod \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.156271 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-svc\") pod \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\" (UID: \"ff3243ad-1c09-4957-8eb6-2413c8132dd4\") " Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.189341 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff3243ad-1c09-4957-8eb6-2413c8132dd4-kube-api-access-xq7x5" (OuterVolumeSpecName: "kube-api-access-xq7x5") pod "ff3243ad-1c09-4957-8eb6-2413c8132dd4" (UID: "ff3243ad-1c09-4957-8eb6-2413c8132dd4"). InnerVolumeSpecName "kube-api-access-xq7x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.202725 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ff3243ad-1c09-4957-8eb6-2413c8132dd4" (UID: "ff3243ad-1c09-4957-8eb6-2413c8132dd4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.210735 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-config" (OuterVolumeSpecName: "config") pod "ff3243ad-1c09-4957-8eb6-2413c8132dd4" (UID: "ff3243ad-1c09-4957-8eb6-2413c8132dd4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.211729 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff3243ad-1c09-4957-8eb6-2413c8132dd4" (UID: "ff3243ad-1c09-4957-8eb6-2413c8132dd4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.221314 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.229718 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff3243ad-1c09-4957-8eb6-2413c8132dd4" (UID: "ff3243ad-1c09-4957-8eb6-2413c8132dd4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.231386 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff3243ad-1c09-4957-8eb6-2413c8132dd4" (UID: "ff3243ad-1c09-4957-8eb6-2413c8132dd4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.259995 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.260029 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq7x5\" (UniqueName: \"kubernetes.io/projected/ff3243ad-1c09-4957-8eb6-2413c8132dd4-kube-api-access-xq7x5\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.260041 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.260049 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.260059 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.260068 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff3243ad-1c09-4957-8eb6-2413c8132dd4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.771758 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" event={"ID":"ff3243ad-1c09-4957-8eb6-2413c8132dd4","Type":"ContainerDied","Data":"a9c1491cea0a47f0c038e6ef3131e2004188f6e737001f40bcbd5837d4f1e5e0"} Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.772073 4935 scope.go:117] "RemoveContainer" containerID="c186f5d7b8259988331ccedc07be2700f0d34e70cc3e5fb3de11ddc492095f7a" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.771818 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd8c5876c-qxqvb" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.777079 4935 generic.go:334] "Generic (PLEG): container finished" podID="30631f10-b2bd-4f3a-9899-54f255c18222" containerID="2d11f160ab6b209c7c31e74f226a72599200f3187ed78af7b2bfbe26c437ee1a" exitCode=0 Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.777148 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" event={"ID":"30631f10-b2bd-4f3a-9899-54f255c18222","Type":"ContainerDied","Data":"2d11f160ab6b209c7c31e74f226a72599200f3187ed78af7b2bfbe26c437ee1a"} Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.799458 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5fwl7" event={"ID":"216a3a9f-2a37-4b71-baf5-0dba7511c4e2","Type":"ContainerStarted","Data":"8b173ce4d56c1415ae5822233d48c1bb3df1e36d7820b9e23201e8af34a167ad"} Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.799506 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5fwl7" event={"ID":"216a3a9f-2a37-4b71-baf5-0dba7511c4e2","Type":"ContainerStarted","Data":"62eeee87fc3e9bf8712e8bae21e840714bdd37de1869d880daf248b9784d3220"} Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.833239 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-5fwl7" podStartSLOduration=2.833219882 podStartE2EDuration="2.833219882s" podCreationTimestamp="2025-10-05 07:12:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:11.823479033 +0000 UTC m=+1165.706105483" watchObservedRunningTime="2025-10-05 07:12:11.833219882 +0000 UTC m=+1165.715846342" Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.888022 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd8c5876c-qxqvb"] Oct 05 07:12:11 crc kubenswrapper[4935]: I1005 07:12:11.888075 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd8c5876c-qxqvb"] Oct 05 07:12:12 crc kubenswrapper[4935]: I1005 07:12:12.804041 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff3243ad-1c09-4957-8eb6-2413c8132dd4" path="/var/lib/kubelet/pods/ff3243ad-1c09-4957-8eb6-2413c8132dd4/volumes" Oct 05 07:12:12 crc kubenswrapper[4935]: I1005 07:12:12.850873 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" event={"ID":"30631f10-b2bd-4f3a-9899-54f255c18222","Type":"ContainerStarted","Data":"9192192a436c9a6d886b109bec87cd56d81b449cd4d976ff7db7443ebc706cfc"} Oct 05 07:12:12 crc kubenswrapper[4935]: I1005 07:12:12.850973 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:12 crc kubenswrapper[4935]: I1005 07:12:12.870801 4935 generic.go:334] "Generic (PLEG): container finished" podID="2da110a9-7934-4fc0-88dc-8025c46413cb" containerID="e61a98d7b38be24992c47d287153ee58d5df1cadd58c8589fb3c0f806844634c" exitCode=0 Oct 05 07:12:12 crc kubenswrapper[4935]: I1005 07:12:12.870862 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mbbgc" event={"ID":"2da110a9-7934-4fc0-88dc-8025c46413cb","Type":"ContainerDied","Data":"e61a98d7b38be24992c47d287153ee58d5df1cadd58c8589fb3c0f806844634c"} Oct 05 07:12:12 crc kubenswrapper[4935]: I1005 07:12:12.914378 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" podStartSLOduration=3.91436188 podStartE2EDuration="3.91436188s" podCreationTimestamp="2025-10-05 07:12:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:12.912229304 +0000 UTC m=+1166.794855764" watchObservedRunningTime="2025-10-05 07:12:12.91436188 +0000 UTC m=+1166.796988340" Oct 05 07:12:13 crc kubenswrapper[4935]: I1005 07:12:13.884798 4935 generic.go:334] "Generic (PLEG): container finished" podID="188231fb-5966-4618-8942-270277384480" containerID="4ac38cab9550b15f03bb7665210b5060057bc358f157b890ebd6d4e945ea3261" exitCode=0 Oct 05 07:12:13 crc kubenswrapper[4935]: I1005 07:12:13.885232 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5pgwt" event={"ID":"188231fb-5966-4618-8942-270277384480","Type":"ContainerDied","Data":"4ac38cab9550b15f03bb7665210b5060057bc358f157b890ebd6d4e945ea3261"} Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.362421 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.372979 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mbbgc" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467126 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-combined-ca-bundle\") pod \"2da110a9-7934-4fc0-88dc-8025c46413cb\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467229 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k8q8\" (UniqueName: \"kubernetes.io/projected/188231fb-5966-4618-8942-270277384480-kube-api-access-7k8q8\") pod \"188231fb-5966-4618-8942-270277384480\" (UID: \"188231fb-5966-4618-8942-270277384480\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467268 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-config-data\") pod \"2da110a9-7934-4fc0-88dc-8025c46413cb\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467307 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-credential-keys\") pod \"188231fb-5966-4618-8942-270277384480\" (UID: \"188231fb-5966-4618-8942-270277384480\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467371 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-db-sync-config-data\") pod \"2da110a9-7934-4fc0-88dc-8025c46413cb\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467403 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-scripts\") pod \"188231fb-5966-4618-8942-270277384480\" (UID: \"188231fb-5966-4618-8942-270277384480\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467443 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-config-data\") pod \"188231fb-5966-4618-8942-270277384480\" (UID: \"188231fb-5966-4618-8942-270277384480\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467548 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-265rb\" (UniqueName: \"kubernetes.io/projected/2da110a9-7934-4fc0-88dc-8025c46413cb-kube-api-access-265rb\") pod \"2da110a9-7934-4fc0-88dc-8025c46413cb\" (UID: \"2da110a9-7934-4fc0-88dc-8025c46413cb\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467587 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-combined-ca-bundle\") pod \"188231fb-5966-4618-8942-270277384480\" (UID: \"188231fb-5966-4618-8942-270277384480\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.467621 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-fernet-keys\") pod \"188231fb-5966-4618-8942-270277384480\" (UID: \"188231fb-5966-4618-8942-270277384480\") " Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.480225 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188231fb-5966-4618-8942-270277384480-kube-api-access-7k8q8" (OuterVolumeSpecName: "kube-api-access-7k8q8") pod "188231fb-5966-4618-8942-270277384480" (UID: "188231fb-5966-4618-8942-270277384480"). InnerVolumeSpecName "kube-api-access-7k8q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.480431 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da110a9-7934-4fc0-88dc-8025c46413cb-kube-api-access-265rb" (OuterVolumeSpecName: "kube-api-access-265rb") pod "2da110a9-7934-4fc0-88dc-8025c46413cb" (UID: "2da110a9-7934-4fc0-88dc-8025c46413cb"). InnerVolumeSpecName "kube-api-access-265rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.481419 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "188231fb-5966-4618-8942-270277384480" (UID: "188231fb-5966-4618-8942-270277384480"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.483345 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-scripts" (OuterVolumeSpecName: "scripts") pod "188231fb-5966-4618-8942-270277384480" (UID: "188231fb-5966-4618-8942-270277384480"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.488147 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2da110a9-7934-4fc0-88dc-8025c46413cb" (UID: "2da110a9-7934-4fc0-88dc-8025c46413cb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.492619 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "188231fb-5966-4618-8942-270277384480" (UID: "188231fb-5966-4618-8942-270277384480"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.496458 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2da110a9-7934-4fc0-88dc-8025c46413cb" (UID: "2da110a9-7934-4fc0-88dc-8025c46413cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.497968 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "188231fb-5966-4618-8942-270277384480" (UID: "188231fb-5966-4618-8942-270277384480"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.502611 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-config-data" (OuterVolumeSpecName: "config-data") pod "188231fb-5966-4618-8942-270277384480" (UID: "188231fb-5966-4618-8942-270277384480"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.522646 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-config-data" (OuterVolumeSpecName: "config-data") pod "2da110a9-7934-4fc0-88dc-8025c46413cb" (UID: "2da110a9-7934-4fc0-88dc-8025c46413cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570227 4935 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570446 4935 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570510 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570562 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570613 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-265rb\" (UniqueName: \"kubernetes.io/projected/2da110a9-7934-4fc0-88dc-8025c46413cb-kube-api-access-265rb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570666 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570720 4935 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/188231fb-5966-4618-8942-270277384480-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570774 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570828 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k8q8\" (UniqueName: \"kubernetes.io/projected/188231fb-5966-4618-8942-270277384480-kube-api-access-7k8q8\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.570880 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da110a9-7934-4fc0-88dc-8025c46413cb-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.902106 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mbbgc" event={"ID":"2da110a9-7934-4fc0-88dc-8025c46413cb","Type":"ContainerDied","Data":"29a22f245ecdc558f6f9a60092d7247ab6ca51e68ea6ec5c3ef1f08b689587f4"} Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.902156 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29a22f245ecdc558f6f9a60092d7247ab6ca51e68ea6ec5c3ef1f08b689587f4" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.902215 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mbbgc" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.910385 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5pgwt" event={"ID":"188231fb-5966-4618-8942-270277384480","Type":"ContainerDied","Data":"02309f5b1bb39825156a62a6640b0bf9ad8b9e57c7f78315ccc8c41bef2f9af0"} Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.910420 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02309f5b1bb39825156a62a6640b0bf9ad8b9e57c7f78315ccc8c41bef2f9af0" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.910438 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5pgwt" Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.977107 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5pgwt"] Oct 05 07:12:15 crc kubenswrapper[4935]: I1005 07:12:15.985979 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5pgwt"] Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.065747 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-k2ff8"] Oct 05 07:12:16 crc kubenswrapper[4935]: E1005 07:12:16.066110 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da110a9-7934-4fc0-88dc-8025c46413cb" containerName="glance-db-sync" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.066126 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da110a9-7934-4fc0-88dc-8025c46413cb" containerName="glance-db-sync" Oct 05 07:12:16 crc kubenswrapper[4935]: E1005 07:12:16.066139 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188231fb-5966-4618-8942-270277384480" containerName="keystone-bootstrap" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.066145 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="188231fb-5966-4618-8942-270277384480" containerName="keystone-bootstrap" Oct 05 07:12:16 crc kubenswrapper[4935]: E1005 07:12:16.066169 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3243ad-1c09-4957-8eb6-2413c8132dd4" containerName="init" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.066176 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3243ad-1c09-4957-8eb6-2413c8132dd4" containerName="init" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.066329 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da110a9-7934-4fc0-88dc-8025c46413cb" containerName="glance-db-sync" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.066340 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff3243ad-1c09-4957-8eb6-2413c8132dd4" containerName="init" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.066363 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="188231fb-5966-4618-8942-270277384480" containerName="keystone-bootstrap" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.067034 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.072083 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.072172 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.072252 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xm842" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.072327 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.088977 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-k2ff8"] Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.186290 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-scripts\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.186612 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54z7w\" (UniqueName: \"kubernetes.io/projected/b520d752-90d4-41a1-a6d0-243e7058c47c-kube-api-access-54z7w\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.186667 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-credential-keys\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.186766 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-config-data\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.186809 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-combined-ca-bundle\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.186833 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-fernet-keys\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.288262 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-config-data\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.288322 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-combined-ca-bundle\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.288349 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-fernet-keys\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.288388 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-scripts\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.288462 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54z7w\" (UniqueName: \"kubernetes.io/projected/b520d752-90d4-41a1-a6d0-243e7058c47c-kube-api-access-54z7w\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.288499 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-credential-keys\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.294596 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-credential-keys\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.295420 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-combined-ca-bundle\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.296190 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-fernet-keys\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.298359 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-scripts\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.310660 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-config-data\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.310793 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54z7w\" (UniqueName: \"kubernetes.io/projected/b520d752-90d4-41a1-a6d0-243e7058c47c-kube-api-access-54z7w\") pod \"keystone-bootstrap-k2ff8\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.389939 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.696174 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dcc79c769-qshg5"] Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.696701 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="dnsmasq-dns" containerID="cri-o://9192192a436c9a6d886b109bec87cd56d81b449cd4d976ff7db7443ebc706cfc" gracePeriod=10 Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.726712 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6795bd975-gs8cj"] Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.739873 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.741396 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-gs8cj"] Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.801583 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188231fb-5966-4618-8942-270277384480" path="/var/lib/kubelet/pods/188231fb-5966-4618-8942-270277384480/volumes" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.906165 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5vn5\" (UniqueName: \"kubernetes.io/projected/00decba6-51fa-4be0-b7be-e74327372758-kube-api-access-j5vn5\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.906420 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-svc\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.906462 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-nb\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.906544 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-sb\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.906565 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-swift-storage-0\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:16 crc kubenswrapper[4935]: I1005 07:12:16.906584 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-config\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.007970 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-svc\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.008054 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-nb\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.008124 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-sb\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.008156 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-swift-storage-0\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.008182 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-config\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.008284 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5vn5\" (UniqueName: \"kubernetes.io/projected/00decba6-51fa-4be0-b7be-e74327372758-kube-api-access-j5vn5\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.009740 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-nb\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.010006 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-config\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.011045 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-sb\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.011240 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-swift-storage-0\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.011960 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-svc\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.034927 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5vn5\" (UniqueName: \"kubernetes.io/projected/00decba6-51fa-4be0-b7be-e74327372758-kube-api-access-j5vn5\") pod \"dnsmasq-dns-6795bd975-gs8cj\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.068318 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.604317 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.605940 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.614689 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.614909 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g86zq" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.615068 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.624152 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.718809 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgmhr\" (UniqueName: \"kubernetes.io/projected/058821c6-d6c2-442a-a5f2-652a60ab79f4-kube-api-access-jgmhr\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.718868 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.718884 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-scripts\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.718937 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-logs\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.718996 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-config-data\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.719235 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.719317 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.820989 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-config-data\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821055 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821071 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821171 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgmhr\" (UniqueName: \"kubernetes.io/projected/058821c6-d6c2-442a-a5f2-652a60ab79f4-kube-api-access-jgmhr\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821194 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821212 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-scripts\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821239 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-logs\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821530 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821635 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.821958 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-logs\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.826208 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.841578 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-scripts\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.843173 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-config-data\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.855557 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgmhr\" (UniqueName: \"kubernetes.io/projected/058821c6-d6c2-442a-a5f2-652a60ab79f4-kube-api-access-jgmhr\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.875431 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.876834 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.883222 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.905282 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.944380 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:12:17 crc kubenswrapper[4935]: I1005 07:12:17.974960 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.006321 4935 generic.go:334] "Generic (PLEG): container finished" podID="30631f10-b2bd-4f3a-9899-54f255c18222" containerID="9192192a436c9a6d886b109bec87cd56d81b449cd4d976ff7db7443ebc706cfc" exitCode=0 Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.006620 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" event={"ID":"30631f10-b2bd-4f3a-9899-54f255c18222","Type":"ContainerDied","Data":"9192192a436c9a6d886b109bec87cd56d81b449cd4d976ff7db7443ebc706cfc"} Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.028161 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.028230 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.028272 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.028307 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.028351 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dwv4\" (UniqueName: \"kubernetes.io/projected/eec81496-90cb-481b-a831-8a6cf858d82a-kube-api-access-8dwv4\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.028407 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.028458 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-logs\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.129988 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-logs\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.130079 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.130108 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.130126 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.130151 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.130180 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dwv4\" (UniqueName: \"kubernetes.io/projected/eec81496-90cb-481b-a831-8a6cf858d82a-kube-api-access-8dwv4\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.130222 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.131162 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-logs\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.131231 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.131259 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.134497 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.134847 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.141511 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.145408 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dwv4\" (UniqueName: \"kubernetes.io/projected/eec81496-90cb-481b-a831-8a6cf858d82a-kube-api-access-8dwv4\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.157544 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:18 crc kubenswrapper[4935]: I1005 07:12:18.315282 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:19 crc kubenswrapper[4935]: I1005 07:12:19.564596 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:19 crc kubenswrapper[4935]: I1005 07:12:19.630454 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:19 crc kubenswrapper[4935]: I1005 07:12:19.748587 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: connect: connection refused" Oct 05 07:12:29 crc kubenswrapper[4935]: I1005 07:12:29.748460 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Oct 05 07:12:29 crc kubenswrapper[4935]: E1005 07:12:29.820391 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:7db347424a8c5998059c5bf84c86a1ef8d582d1ffe39f4887551f2ac85a4915f" Oct 05 07:12:29 crc kubenswrapper[4935]: E1005 07:12:29.820740 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:7db347424a8c5998059c5bf84c86a1ef8d582d1ffe39f4887551f2ac85a4915f,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-skwwn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-nl9vd_openstack(2ef780aa-5f7e-4fa2-a019-6ee2c3993692): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:12:29 crc kubenswrapper[4935]: E1005 07:12:29.822930 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-nl9vd" podUID="2ef780aa-5f7e-4fa2-a019-6ee2c3993692" Oct 05 07:12:29 crc kubenswrapper[4935]: I1005 07:12:29.907556 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.037328 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-svc\") pod \"30631f10-b2bd-4f3a-9899-54f255c18222\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.037376 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-swift-storage-0\") pod \"30631f10-b2bd-4f3a-9899-54f255c18222\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.037427 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-config\") pod \"30631f10-b2bd-4f3a-9899-54f255c18222\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.037522 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-sb\") pod \"30631f10-b2bd-4f3a-9899-54f255c18222\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.037543 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-nb\") pod \"30631f10-b2bd-4f3a-9899-54f255c18222\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.037603 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdxc7\" (UniqueName: \"kubernetes.io/projected/30631f10-b2bd-4f3a-9899-54f255c18222-kube-api-access-xdxc7\") pod \"30631f10-b2bd-4f3a-9899-54f255c18222\" (UID: \"30631f10-b2bd-4f3a-9899-54f255c18222\") " Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.041686 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30631f10-b2bd-4f3a-9899-54f255c18222-kube-api-access-xdxc7" (OuterVolumeSpecName: "kube-api-access-xdxc7") pod "30631f10-b2bd-4f3a-9899-54f255c18222" (UID: "30631f10-b2bd-4f3a-9899-54f255c18222"). InnerVolumeSpecName "kube-api-access-xdxc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.084499 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "30631f10-b2bd-4f3a-9899-54f255c18222" (UID: "30631f10-b2bd-4f3a-9899-54f255c18222"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.087073 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "30631f10-b2bd-4f3a-9899-54f255c18222" (UID: "30631f10-b2bd-4f3a-9899-54f255c18222"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.089841 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "30631f10-b2bd-4f3a-9899-54f255c18222" (UID: "30631f10-b2bd-4f3a-9899-54f255c18222"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.090435 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "30631f10-b2bd-4f3a-9899-54f255c18222" (UID: "30631f10-b2bd-4f3a-9899-54f255c18222"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.104978 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-config" (OuterVolumeSpecName: "config") pod "30631f10-b2bd-4f3a-9899-54f255c18222" (UID: "30631f10-b2bd-4f3a-9899-54f255c18222"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.127476 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.128121 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" event={"ID":"30631f10-b2bd-4f3a-9899-54f255c18222","Type":"ContainerDied","Data":"cfcfeca7fe30a1f4f41a4453250d850c8fb34760875eda0e3e46a93cdd3e9446"} Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.128180 4935 scope.go:117] "RemoveContainer" containerID="9192192a436c9a6d886b109bec87cd56d81b449cd4d976ff7db7443ebc706cfc" Oct 05 07:12:30 crc kubenswrapper[4935]: E1005 07:12:30.130053 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:7db347424a8c5998059c5bf84c86a1ef8d582d1ffe39f4887551f2ac85a4915f\\\"\"" pod="openstack/barbican-db-sync-nl9vd" podUID="2ef780aa-5f7e-4fa2-a019-6ee2c3993692" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.139945 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.139981 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.139995 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdxc7\" (UniqueName: \"kubernetes.io/projected/30631f10-b2bd-4f3a-9899-54f255c18222-kube-api-access-xdxc7\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.140010 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.140023 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.140034 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30631f10-b2bd-4f3a-9899-54f255c18222-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.182733 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dcc79c769-qshg5"] Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.190451 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dcc79c769-qshg5"] Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.785508 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" path="/var/lib/kubelet/pods/30631f10-b2bd-4f3a-9899-54f255c18222/volumes" Oct 05 07:12:30 crc kubenswrapper[4935]: I1005 07:12:30.865061 4935 scope.go:117] "RemoveContainer" containerID="2d11f160ab6b209c7c31e74f226a72599200f3187ed78af7b2bfbe26c437ee1a" Oct 05 07:12:30 crc kubenswrapper[4935]: E1005 07:12:30.882157 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213" Oct 05 07:12:30 crc kubenswrapper[4935]: E1005 07:12:30.883384 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z26l2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-n9vr5_openstack(e9ba8cbb-45b9-49e5-b5a0-9277b724b607): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:12:30 crc kubenswrapper[4935]: E1005 07:12:30.884686 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-n9vr5" podUID="e9ba8cbb-45b9-49e5-b5a0-9277b724b607" Oct 05 07:12:31 crc kubenswrapper[4935]: E1005 07:12:31.136777 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213\\\"\"" pod="openstack/cinder-db-sync-n9vr5" podUID="e9ba8cbb-45b9-49e5-b5a0-9277b724b607" Oct 05 07:12:31 crc kubenswrapper[4935]: I1005 07:12:31.254851 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-k2ff8"] Oct 05 07:12:31 crc kubenswrapper[4935]: I1005 07:12:31.346802 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-gs8cj"] Oct 05 07:12:31 crc kubenswrapper[4935]: W1005 07:12:31.650816 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeec81496_90cb_481b_a831_8a6cf858d82a.slice/crio-d0f5615b3c935cdcb1a530db69356cb66b11b2bd9623ef00385e0509a225627e WatchSource:0}: Error finding container d0f5615b3c935cdcb1a530db69356cb66b11b2bd9623ef00385e0509a225627e: Status 404 returned error can't find the container with id d0f5615b3c935cdcb1a530db69356cb66b11b2bd9623ef00385e0509a225627e Oct 05 07:12:31 crc kubenswrapper[4935]: I1005 07:12:31.654972 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.147085 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vfdz" event={"ID":"6e808378-4679-42ed-a47c-033e1edfb362","Type":"ContainerStarted","Data":"dae10e0c5cc975b8622037342618fdb57f961bf238a26977922a4a5038cc681f"} Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.149078 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"eec81496-90cb-481b-a831-8a6cf858d82a","Type":"ContainerStarted","Data":"d0f5615b3c935cdcb1a530db69356cb66b11b2bd9623ef00385e0509a225627e"} Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.151577 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerStarted","Data":"7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde"} Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.153087 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k2ff8" event={"ID":"b520d752-90d4-41a1-a6d0-243e7058c47c","Type":"ContainerStarted","Data":"7415798dcf5ca1b35b5ba66c1a74be46595f88e290eaa7b69d58158f804e237f"} Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.153115 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k2ff8" event={"ID":"b520d752-90d4-41a1-a6d0-243e7058c47c","Type":"ContainerStarted","Data":"38c577023ff6f5e54a282572ca4bb1643eb7324ba233f89f8f161a0a79a8a8c1"} Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.156175 4935 generic.go:334] "Generic (PLEG): container finished" podID="00decba6-51fa-4be0-b7be-e74327372758" containerID="056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa" exitCode=0 Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.156226 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" event={"ID":"00decba6-51fa-4be0-b7be-e74327372758","Type":"ContainerDied","Data":"056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa"} Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.156254 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" event={"ID":"00decba6-51fa-4be0-b7be-e74327372758","Type":"ContainerStarted","Data":"be1a5634d957122e1b48bea99487e888a3563a3c1a8d7f9d78e200947a232565"} Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.180190 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-2vfdz" podStartSLOduration=3.9295819229999998 podStartE2EDuration="23.180167044s" podCreationTimestamp="2025-10-05 07:12:09 +0000 UTC" firstStartedPulling="2025-10-05 07:12:10.564578949 +0000 UTC m=+1164.447205399" lastFinishedPulling="2025-10-05 07:12:29.81516406 +0000 UTC m=+1183.697790520" observedRunningTime="2025-10-05 07:12:32.173266271 +0000 UTC m=+1186.055892741" watchObservedRunningTime="2025-10-05 07:12:32.180167044 +0000 UTC m=+1186.062793504" Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.219972 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-k2ff8" podStartSLOduration=16.219951273 podStartE2EDuration="16.219951273s" podCreationTimestamp="2025-10-05 07:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:32.213336157 +0000 UTC m=+1186.095962627" watchObservedRunningTime="2025-10-05 07:12:32.219951273 +0000 UTC m=+1186.102577743" Oct 05 07:12:32 crc kubenswrapper[4935]: I1005 07:12:32.488571 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:32 crc kubenswrapper[4935]: W1005 07:12:32.494942 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod058821c6_d6c2_442a_a5f2_652a60ab79f4.slice/crio-bc474477e05cdf44c09b5908a24a67d949da505c9346b6e4efb4f17014dba5b7 WatchSource:0}: Error finding container bc474477e05cdf44c09b5908a24a67d949da505c9346b6e4efb4f17014dba5b7: Status 404 returned error can't find the container with id bc474477e05cdf44c09b5908a24a67d949da505c9346b6e4efb4f17014dba5b7 Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.167448 4935 generic.go:334] "Generic (PLEG): container finished" podID="216a3a9f-2a37-4b71-baf5-0dba7511c4e2" containerID="8b173ce4d56c1415ae5822233d48c1bb3df1e36d7820b9e23201e8af34a167ad" exitCode=0 Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.167541 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5fwl7" event={"ID":"216a3a9f-2a37-4b71-baf5-0dba7511c4e2","Type":"ContainerDied","Data":"8b173ce4d56c1415ae5822233d48c1bb3df1e36d7820b9e23201e8af34a167ad"} Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.170935 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" event={"ID":"00decba6-51fa-4be0-b7be-e74327372758","Type":"ContainerStarted","Data":"cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f"} Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.171353 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.172489 4935 generic.go:334] "Generic (PLEG): container finished" podID="6e808378-4679-42ed-a47c-033e1edfb362" containerID="dae10e0c5cc975b8622037342618fdb57f961bf238a26977922a4a5038cc681f" exitCode=0 Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.172543 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vfdz" event={"ID":"6e808378-4679-42ed-a47c-033e1edfb362","Type":"ContainerDied","Data":"dae10e0c5cc975b8622037342618fdb57f961bf238a26977922a4a5038cc681f"} Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.178763 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"eec81496-90cb-481b-a831-8a6cf858d82a","Type":"ContainerStarted","Data":"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f"} Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.178818 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"eec81496-90cb-481b-a831-8a6cf858d82a","Type":"ContainerStarted","Data":"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88"} Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.179008 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" containerName="glance-log" containerID="cri-o://8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88" gracePeriod=30 Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.179301 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" containerName="glance-httpd" containerID="cri-o://cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f" gracePeriod=30 Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.197843 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"058821c6-d6c2-442a-a5f2-652a60ab79f4","Type":"ContainerStarted","Data":"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1"} Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.198023 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"058821c6-d6c2-442a-a5f2-652a60ab79f4","Type":"ContainerStarted","Data":"bc474477e05cdf44c09b5908a24a67d949da505c9346b6e4efb4f17014dba5b7"} Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.227022 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.226999047 podStartE2EDuration="17.226999047s" podCreationTimestamp="2025-10-05 07:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:33.210089477 +0000 UTC m=+1187.092716047" watchObservedRunningTime="2025-10-05 07:12:33.226999047 +0000 UTC m=+1187.109625517" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.255989 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" podStartSLOduration=17.255971418 podStartE2EDuration="17.255971418s" podCreationTimestamp="2025-10-05 07:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:33.251253833 +0000 UTC m=+1187.133880293" watchObservedRunningTime="2025-10-05 07:12:33.255971418 +0000 UTC m=+1187.138597868" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.768012 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.835049 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-scripts\") pod \"eec81496-90cb-481b-a831-8a6cf858d82a\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.835088 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"eec81496-90cb-481b-a831-8a6cf858d82a\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.835174 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-config-data\") pod \"eec81496-90cb-481b-a831-8a6cf858d82a\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.835214 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-httpd-run\") pod \"eec81496-90cb-481b-a831-8a6cf858d82a\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.835266 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dwv4\" (UniqueName: \"kubernetes.io/projected/eec81496-90cb-481b-a831-8a6cf858d82a-kube-api-access-8dwv4\") pod \"eec81496-90cb-481b-a831-8a6cf858d82a\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.835303 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-combined-ca-bundle\") pod \"eec81496-90cb-481b-a831-8a6cf858d82a\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.835327 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-logs\") pod \"eec81496-90cb-481b-a831-8a6cf858d82a\" (UID: \"eec81496-90cb-481b-a831-8a6cf858d82a\") " Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.841202 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "eec81496-90cb-481b-a831-8a6cf858d82a" (UID: "eec81496-90cb-481b-a831-8a6cf858d82a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.842090 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-logs" (OuterVolumeSpecName: "logs") pod "eec81496-90cb-481b-a831-8a6cf858d82a" (UID: "eec81496-90cb-481b-a831-8a6cf858d82a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.843264 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-scripts" (OuterVolumeSpecName: "scripts") pod "eec81496-90cb-481b-a831-8a6cf858d82a" (UID: "eec81496-90cb-481b-a831-8a6cf858d82a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.844868 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "eec81496-90cb-481b-a831-8a6cf858d82a" (UID: "eec81496-90cb-481b-a831-8a6cf858d82a"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.847261 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eec81496-90cb-481b-a831-8a6cf858d82a-kube-api-access-8dwv4" (OuterVolumeSpecName: "kube-api-access-8dwv4") pod "eec81496-90cb-481b-a831-8a6cf858d82a" (UID: "eec81496-90cb-481b-a831-8a6cf858d82a"). InnerVolumeSpecName "kube-api-access-8dwv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.882133 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eec81496-90cb-481b-a831-8a6cf858d82a" (UID: "eec81496-90cb-481b-a831-8a6cf858d82a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.899398 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-config-data" (OuterVolumeSpecName: "config-data") pod "eec81496-90cb-481b-a831-8a6cf858d82a" (UID: "eec81496-90cb-481b-a831-8a6cf858d82a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.938291 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dwv4\" (UniqueName: \"kubernetes.io/projected/eec81496-90cb-481b-a831-8a6cf858d82a-kube-api-access-8dwv4\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.938331 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.938348 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.938362 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.938402 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.938419 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec81496-90cb-481b-a831-8a6cf858d82a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.938432 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec81496-90cb-481b-a831-8a6cf858d82a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:33 crc kubenswrapper[4935]: I1005 07:12:33.966524 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.039813 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.228472 4935 generic.go:334] "Generic (PLEG): container finished" podID="eec81496-90cb-481b-a831-8a6cf858d82a" containerID="cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f" exitCode=0 Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.228530 4935 generic.go:334] "Generic (PLEG): container finished" podID="eec81496-90cb-481b-a831-8a6cf858d82a" containerID="8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88" exitCode=143 Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.228601 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"eec81496-90cb-481b-a831-8a6cf858d82a","Type":"ContainerDied","Data":"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f"} Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.228646 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"eec81496-90cb-481b-a831-8a6cf858d82a","Type":"ContainerDied","Data":"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88"} Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.228667 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"eec81496-90cb-481b-a831-8a6cf858d82a","Type":"ContainerDied","Data":"d0f5615b3c935cdcb1a530db69356cb66b11b2bd9623ef00385e0509a225627e"} Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.228693 4935 scope.go:117] "RemoveContainer" containerID="cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.228878 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.234732 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerStarted","Data":"52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06"} Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.237088 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"058821c6-d6c2-442a-a5f2-652a60ab79f4","Type":"ContainerStarted","Data":"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1"} Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.237486 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerName="glance-log" containerID="cri-o://c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1" gracePeriod=30 Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.237927 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerName="glance-httpd" containerID="cri-o://251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1" gracePeriod=30 Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.243982 4935 generic.go:334] "Generic (PLEG): container finished" podID="b520d752-90d4-41a1-a6d0-243e7058c47c" containerID="7415798dcf5ca1b35b5ba66c1a74be46595f88e290eaa7b69d58158f804e237f" exitCode=0 Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.244835 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k2ff8" event={"ID":"b520d752-90d4-41a1-a6d0-243e7058c47c","Type":"ContainerDied","Data":"7415798dcf5ca1b35b5ba66c1a74be46595f88e290eaa7b69d58158f804e237f"} Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.276264 4935 scope.go:117] "RemoveContainer" containerID="8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.305715 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=18.305694998 podStartE2EDuration="18.305694998s" podCreationTimestamp="2025-10-05 07:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:34.260360392 +0000 UTC m=+1188.142986862" watchObservedRunningTime="2025-10-05 07:12:34.305694998 +0000 UTC m=+1188.188321458" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.318161 4935 scope.go:117] "RemoveContainer" containerID="cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.318577 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:34 crc kubenswrapper[4935]: E1005 07:12:34.318966 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f\": container with ID starting with cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f not found: ID does not exist" containerID="cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.318997 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f"} err="failed to get container status \"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f\": rpc error: code = NotFound desc = could not find container \"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f\": container with ID starting with cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f not found: ID does not exist" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.319018 4935 scope.go:117] "RemoveContainer" containerID="8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88" Oct 05 07:12:34 crc kubenswrapper[4935]: E1005 07:12:34.319269 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88\": container with ID starting with 8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88 not found: ID does not exist" containerID="8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.319293 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88"} err="failed to get container status \"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88\": rpc error: code = NotFound desc = could not find container \"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88\": container with ID starting with 8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88 not found: ID does not exist" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.319307 4935 scope.go:117] "RemoveContainer" containerID="cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.319463 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f"} err="failed to get container status \"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f\": rpc error: code = NotFound desc = could not find container \"cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f\": container with ID starting with cfd65f94a7d8451bb54413dbf3c3ccc2b4f0ab8d2f4dd5e28976b066dc07361f not found: ID does not exist" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.319482 4935 scope.go:117] "RemoveContainer" containerID="8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.319636 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88"} err="failed to get container status \"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88\": rpc error: code = NotFound desc = could not find container \"8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88\": container with ID starting with 8c9ce3adb5cc63c1bad70a736294739d37189105a100c1226f65634d25fe9a88 not found: ID does not exist" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.325989 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.335637 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:34 crc kubenswrapper[4935]: E1005 07:12:34.336141 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" containerName="glance-httpd" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.336166 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" containerName="glance-httpd" Oct 05 07:12:34 crc kubenswrapper[4935]: E1005 07:12:34.336188 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="init" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.336197 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="init" Oct 05 07:12:34 crc kubenswrapper[4935]: E1005 07:12:34.336226 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" containerName="glance-log" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.336234 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" containerName="glance-log" Oct 05 07:12:34 crc kubenswrapper[4935]: E1005 07:12:34.336246 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="dnsmasq-dns" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.336254 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="dnsmasq-dns" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.336500 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" containerName="glance-log" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.336745 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" containerName="glance-httpd" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.336763 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="dnsmasq-dns" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.338040 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.350317 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.350553 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.354176 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.451291 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.451345 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-logs\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.451397 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.451546 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7jcq\" (UniqueName: \"kubernetes.io/projected/980fa02a-09b0-42cc-8028-51b588427002-kube-api-access-b7jcq\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.451646 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-config-data\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.451695 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.452024 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-scripts\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.452075 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.555342 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.555388 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-logs\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.555430 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.555484 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7jcq\" (UniqueName: \"kubernetes.io/projected/980fa02a-09b0-42cc-8028-51b588427002-kube-api-access-b7jcq\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.555506 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-config-data\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.555530 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.555580 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-scripts\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.555610 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.556039 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.557113 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-logs\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.557368 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.563598 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-scripts\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.563796 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.564125 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.569828 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-config-data\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.571617 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7jcq\" (UniqueName: \"kubernetes.io/projected/980fa02a-09b0-42cc-8028-51b588427002-kube-api-access-b7jcq\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.586664 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.610009 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.655236 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.673458 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.750055 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-dcc79c769-qshg5" podUID="30631f10-b2bd-4f3a-9899-54f255c18222" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.757663 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-combined-ca-bundle\") pod \"6e808378-4679-42ed-a47c-033e1edfb362\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.757697 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e808378-4679-42ed-a47c-033e1edfb362-logs\") pod \"6e808378-4679-42ed-a47c-033e1edfb362\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.757804 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-scripts\") pod \"6e808378-4679-42ed-a47c-033e1edfb362\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.757830 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-config-data\") pod \"6e808378-4679-42ed-a47c-033e1edfb362\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.757906 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z22w8\" (UniqueName: \"kubernetes.io/projected/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-kube-api-access-z22w8\") pod \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.757930 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-combined-ca-bundle\") pod \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.757954 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-config\") pod \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\" (UID: \"216a3a9f-2a37-4b71-baf5-0dba7511c4e2\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.758000 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsjtt\" (UniqueName: \"kubernetes.io/projected/6e808378-4679-42ed-a47c-033e1edfb362-kube-api-access-bsjtt\") pod \"6e808378-4679-42ed-a47c-033e1edfb362\" (UID: \"6e808378-4679-42ed-a47c-033e1edfb362\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.759246 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e808378-4679-42ed-a47c-033e1edfb362-logs" (OuterVolumeSpecName: "logs") pod "6e808378-4679-42ed-a47c-033e1edfb362" (UID: "6e808378-4679-42ed-a47c-033e1edfb362"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.762498 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e808378-4679-42ed-a47c-033e1edfb362-kube-api-access-bsjtt" (OuterVolumeSpecName: "kube-api-access-bsjtt") pod "6e808378-4679-42ed-a47c-033e1edfb362" (UID: "6e808378-4679-42ed-a47c-033e1edfb362"). InnerVolumeSpecName "kube-api-access-bsjtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.763516 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-scripts" (OuterVolumeSpecName: "scripts") pod "6e808378-4679-42ed-a47c-033e1edfb362" (UID: "6e808378-4679-42ed-a47c-033e1edfb362"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.764068 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-kube-api-access-z22w8" (OuterVolumeSpecName: "kube-api-access-z22w8") pod "216a3a9f-2a37-4b71-baf5-0dba7511c4e2" (UID: "216a3a9f-2a37-4b71-baf5-0dba7511c4e2"). InnerVolumeSpecName "kube-api-access-z22w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.789120 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "216a3a9f-2a37-4b71-baf5-0dba7511c4e2" (UID: "216a3a9f-2a37-4b71-baf5-0dba7511c4e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.790512 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-config" (OuterVolumeSpecName: "config") pod "216a3a9f-2a37-4b71-baf5-0dba7511c4e2" (UID: "216a3a9f-2a37-4b71-baf5-0dba7511c4e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.805399 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eec81496-90cb-481b-a831-8a6cf858d82a" path="/var/lib/kubelet/pods/eec81496-90cb-481b-a831-8a6cf858d82a/volumes" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.822589 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.823144 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e808378-4679-42ed-a47c-033e1edfb362" (UID: "6e808378-4679-42ed-a47c-033e1edfb362"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.833224 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-config-data" (OuterVolumeSpecName: "config-data") pod "6e808378-4679-42ed-a47c-033e1edfb362" (UID: "6e808378-4679-42ed-a47c-033e1edfb362"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.860463 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z22w8\" (UniqueName: \"kubernetes.io/projected/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-kube-api-access-z22w8\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.860498 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.860509 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/216a3a9f-2a37-4b71-baf5-0dba7511c4e2-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.860519 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsjtt\" (UniqueName: \"kubernetes.io/projected/6e808378-4679-42ed-a47c-033e1edfb362-kube-api-access-bsjtt\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.860529 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.861074 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e808378-4679-42ed-a47c-033e1edfb362-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.861092 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.861102 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e808378-4679-42ed-a47c-033e1edfb362-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.961705 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-combined-ca-bundle\") pod \"058821c6-d6c2-442a-a5f2-652a60ab79f4\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.961743 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-scripts\") pod \"058821c6-d6c2-442a-a5f2-652a60ab79f4\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.961849 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"058821c6-d6c2-442a-a5f2-652a60ab79f4\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.961873 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-config-data\") pod \"058821c6-d6c2-442a-a5f2-652a60ab79f4\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.961917 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-httpd-run\") pod \"058821c6-d6c2-442a-a5f2-652a60ab79f4\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.961955 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgmhr\" (UniqueName: \"kubernetes.io/projected/058821c6-d6c2-442a-a5f2-652a60ab79f4-kube-api-access-jgmhr\") pod \"058821c6-d6c2-442a-a5f2-652a60ab79f4\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.962005 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-logs\") pod \"058821c6-d6c2-442a-a5f2-652a60ab79f4\" (UID: \"058821c6-d6c2-442a-a5f2-652a60ab79f4\") " Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.962818 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "058821c6-d6c2-442a-a5f2-652a60ab79f4" (UID: "058821c6-d6c2-442a-a5f2-652a60ab79f4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.962838 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-logs" (OuterVolumeSpecName: "logs") pod "058821c6-d6c2-442a-a5f2-652a60ab79f4" (UID: "058821c6-d6c2-442a-a5f2-652a60ab79f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.966805 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/058821c6-d6c2-442a-a5f2-652a60ab79f4-kube-api-access-jgmhr" (OuterVolumeSpecName: "kube-api-access-jgmhr") pod "058821c6-d6c2-442a-a5f2-652a60ab79f4" (UID: "058821c6-d6c2-442a-a5f2-652a60ab79f4"). InnerVolumeSpecName "kube-api-access-jgmhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.968036 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-scripts" (OuterVolumeSpecName: "scripts") pod "058821c6-d6c2-442a-a5f2-652a60ab79f4" (UID: "058821c6-d6c2-442a-a5f2-652a60ab79f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:34 crc kubenswrapper[4935]: I1005 07:12:34.972026 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "058821c6-d6c2-442a-a5f2-652a60ab79f4" (UID: "058821c6-d6c2-442a-a5f2-652a60ab79f4"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.009985 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "058821c6-d6c2-442a-a5f2-652a60ab79f4" (UID: "058821c6-d6c2-442a-a5f2-652a60ab79f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.024764 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-config-data" (OuterVolumeSpecName: "config-data") pod "058821c6-d6c2-442a-a5f2-652a60ab79f4" (UID: "058821c6-d6c2-442a-a5f2-652a60ab79f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.063572 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.063608 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.063641 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.063650 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/058821c6-d6c2-442a-a5f2-652a60ab79f4-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.063659 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.063667 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgmhr\" (UniqueName: \"kubernetes.io/projected/058821c6-d6c2-442a-a5f2-652a60ab79f4-kube-api-access-jgmhr\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.063680 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/058821c6-d6c2-442a-a5f2-652a60ab79f4-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.083037 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.166126 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.192548 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:12:35 crc kubenswrapper[4935]: W1005 07:12:35.225251 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod980fa02a_09b0_42cc_8028_51b588427002.slice/crio-19f1e4a6406ea7af77cb699b64a82111391d5c4cdb9f6da517169509001317e7 WatchSource:0}: Error finding container 19f1e4a6406ea7af77cb699b64a82111391d5c4cdb9f6da517169509001317e7: Status 404 returned error can't find the container with id 19f1e4a6406ea7af77cb699b64a82111391d5c4cdb9f6da517169509001317e7 Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.259936 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vfdz" event={"ID":"6e808378-4679-42ed-a47c-033e1edfb362","Type":"ContainerDied","Data":"096f3674640cbb85a5df9f4b6849b5a89a5b0ddfcda84e8ff239d0fb6d4f32d7"} Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.259983 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="096f3674640cbb85a5df9f4b6849b5a89a5b0ddfcda84e8ff239d0fb6d4f32d7" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.260008 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vfdz" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.262856 4935 generic.go:334] "Generic (PLEG): container finished" podID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerID="251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1" exitCode=0 Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.262880 4935 generic.go:334] "Generic (PLEG): container finished" podID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerID="c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1" exitCode=143 Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.262935 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"058821c6-d6c2-442a-a5f2-652a60ab79f4","Type":"ContainerDied","Data":"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1"} Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.262960 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"058821c6-d6c2-442a-a5f2-652a60ab79f4","Type":"ContainerDied","Data":"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1"} Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.262974 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"058821c6-d6c2-442a-a5f2-652a60ab79f4","Type":"ContainerDied","Data":"bc474477e05cdf44c09b5908a24a67d949da505c9346b6e4efb4f17014dba5b7"} Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.262994 4935 scope.go:117] "RemoveContainer" containerID="251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.263114 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.282560 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"980fa02a-09b0-42cc-8028-51b588427002","Type":"ContainerStarted","Data":"19f1e4a6406ea7af77cb699b64a82111391d5c4cdb9f6da517169509001317e7"} Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.289164 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5fwl7" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.289221 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5fwl7" event={"ID":"216a3a9f-2a37-4b71-baf5-0dba7511c4e2","Type":"ContainerDied","Data":"62eeee87fc3e9bf8712e8bae21e840714bdd37de1869d880daf248b9784d3220"} Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.289259 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62eeee87fc3e9bf8712e8bae21e840714bdd37de1869d880daf248b9784d3220" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.291688 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6dbdfd7554-jrf6q"] Oct 05 07:12:35 crc kubenswrapper[4935]: E1005 07:12:35.297018 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerName="glance-log" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.297034 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerName="glance-log" Oct 05 07:12:35 crc kubenswrapper[4935]: E1005 07:12:35.297045 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="216a3a9f-2a37-4b71-baf5-0dba7511c4e2" containerName="neutron-db-sync" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.297052 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="216a3a9f-2a37-4b71-baf5-0dba7511c4e2" containerName="neutron-db-sync" Oct 05 07:12:35 crc kubenswrapper[4935]: E1005 07:12:35.297066 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerName="glance-httpd" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.297072 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerName="glance-httpd" Oct 05 07:12:35 crc kubenswrapper[4935]: E1005 07:12:35.297081 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e808378-4679-42ed-a47c-033e1edfb362" containerName="placement-db-sync" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.297087 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e808378-4679-42ed-a47c-033e1edfb362" containerName="placement-db-sync" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.297260 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="216a3a9f-2a37-4b71-baf5-0dba7511c4e2" containerName="neutron-db-sync" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.297284 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerName="glance-httpd" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.297295 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" containerName="glance-log" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.297306 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e808378-4679-42ed-a47c-033e1edfb362" containerName="placement-db-sync" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.298403 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.302453 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.302467 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.304339 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.304467 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.304610 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vbqqd" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.330668 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6dbdfd7554-jrf6q"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.347230 4935 scope.go:117] "RemoveContainer" containerID="c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.391345 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-internal-tls-certs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.396015 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-public-tls-certs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.396197 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-logs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.396301 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.396414 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.396572 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-combined-ca-bundle\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.396638 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp87c\" (UniqueName: \"kubernetes.io/projected/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-kube-api-access-gp87c\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.394528 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.411968 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.413066 4935 scope.go:117] "RemoveContainer" containerID="251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1" Oct 05 07:12:35 crc kubenswrapper[4935]: E1005 07:12:35.421825 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1\": container with ID starting with 251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1 not found: ID does not exist" containerID="251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.421872 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1"} err="failed to get container status \"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1\": rpc error: code = NotFound desc = could not find container \"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1\": container with ID starting with 251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1 not found: ID does not exist" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.421923 4935 scope.go:117] "RemoveContainer" containerID="c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1" Oct 05 07:12:35 crc kubenswrapper[4935]: E1005 07:12:35.426337 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1\": container with ID starting with c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1 not found: ID does not exist" containerID="c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.426374 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1"} err="failed to get container status \"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1\": rpc error: code = NotFound desc = could not find container \"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1\": container with ID starting with c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1 not found: ID does not exist" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.426396 4935 scope.go:117] "RemoveContainer" containerID="251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.426936 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1"} err="failed to get container status \"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1\": rpc error: code = NotFound desc = could not find container \"251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1\": container with ID starting with 251bdad89885054be8d58dc0edc7bee43d9620ec350a9f2c15e34426f6d14ee1 not found: ID does not exist" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.427012 4935 scope.go:117] "RemoveContainer" containerID="c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.427340 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1"} err="failed to get container status \"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1\": rpc error: code = NotFound desc = could not find container \"c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1\": container with ID starting with c4a509c57da4018e8f36e3fae63590b5f82e8ffe4e5d7f21f08bee6c3bed85b1 not found: ID does not exist" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.431711 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.433174 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.443831 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.443913 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.480470 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.506413 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-public-tls-certs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.506503 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-logs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.506541 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.506584 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.506668 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-combined-ca-bundle\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.506689 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp87c\" (UniqueName: \"kubernetes.io/projected/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-kube-api-access-gp87c\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.506728 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-internal-tls-certs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.515186 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-gs8cj"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.515705 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" podUID="00decba6-51fa-4be0-b7be-e74327372758" containerName="dnsmasq-dns" containerID="cri-o://cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f" gracePeriod=10 Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.520788 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-internal-tls-certs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.524791 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-logs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.529477 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.532249 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-combined-ca-bundle\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.533144 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-public-tls-certs\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.533771 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.569239 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp87c\" (UniqueName: \"kubernetes.io/projected/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-kube-api-access-gp87c\") pod \"placement-6dbdfd7554-jrf6q\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.601162 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc97d9d55-72qqm"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.605198 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.630634 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.630803 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.631667 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.631766 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.631808 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-logs\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.631983 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.632030 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knfb6\" (UniqueName: \"kubernetes.io/projected/b885bd46-8ccc-47d6-8622-c4175e2267e5-kube-api-access-knfb6\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.632060 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.640010 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-56b79f594d-q76hh"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.646466 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.662849 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.667943 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc97d9d55-72qqm"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.668664 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.668813 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8g6z4" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.668939 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.670257 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.683313 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56b79f594d-q76hh"] Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734513 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-config\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734568 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-svc\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734601 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734639 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-swift-storage-0\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734670 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734694 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734712 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cscw6\" (UniqueName: \"kubernetes.io/projected/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-kube-api-access-cscw6\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734734 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734764 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734782 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-logs\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734878 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734914 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734941 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knfb6\" (UniqueName: \"kubernetes.io/projected/b885bd46-8ccc-47d6-8622-c4175e2267e5-kube-api-access-knfb6\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.734960 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.735870 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.737026 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.739345 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-logs\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.740144 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.741585 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.743313 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.750939 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.761481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knfb6\" (UniqueName: \"kubernetes.io/projected/b885bd46-8ccc-47d6-8622-c4175e2267e5-kube-api-access-knfb6\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.766739 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " pod="openstack/glance-default-external-api-0" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.836944 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-combined-ca-bundle\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837351 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837378 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-httpd-config\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837420 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-config\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837441 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-svc\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837478 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-config\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837502 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-swift-storage-0\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837530 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837547 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cscw6\" (UniqueName: \"kubernetes.io/projected/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-kube-api-access-cscw6\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837566 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrthr\" (UniqueName: \"kubernetes.io/projected/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-kube-api-access-xrthr\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.837597 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-ovndb-tls-certs\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.838761 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-config\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.840820 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-swift-storage-0\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.842481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.842849 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-svc\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.843314 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.854916 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.870702 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cscw6\" (UniqueName: \"kubernetes.io/projected/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-kube-api-access-cscw6\") pod \"dnsmasq-dns-6bc97d9d55-72qqm\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.939269 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54z7w\" (UniqueName: \"kubernetes.io/projected/b520d752-90d4-41a1-a6d0-243e7058c47c-kube-api-access-54z7w\") pod \"b520d752-90d4-41a1-a6d0-243e7058c47c\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.939314 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-scripts\") pod \"b520d752-90d4-41a1-a6d0-243e7058c47c\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.939403 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-combined-ca-bundle\") pod \"b520d752-90d4-41a1-a6d0-243e7058c47c\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.939485 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-credential-keys\") pod \"b520d752-90d4-41a1-a6d0-243e7058c47c\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.939502 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-fernet-keys\") pod \"b520d752-90d4-41a1-a6d0-243e7058c47c\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.939750 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-config-data\") pod \"b520d752-90d4-41a1-a6d0-243e7058c47c\" (UID: \"b520d752-90d4-41a1-a6d0-243e7058c47c\") " Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.945970 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b520d752-90d4-41a1-a6d0-243e7058c47c" (UID: "b520d752-90d4-41a1-a6d0-243e7058c47c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.946099 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b520d752-90d4-41a1-a6d0-243e7058c47c-kube-api-access-54z7w" (OuterVolumeSpecName: "kube-api-access-54z7w") pod "b520d752-90d4-41a1-a6d0-243e7058c47c" (UID: "b520d752-90d4-41a1-a6d0-243e7058c47c"). InnerVolumeSpecName "kube-api-access-54z7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.946270 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-scripts" (OuterVolumeSpecName: "scripts") pod "b520d752-90d4-41a1-a6d0-243e7058c47c" (UID: "b520d752-90d4-41a1-a6d0-243e7058c47c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.953000 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b520d752-90d4-41a1-a6d0-243e7058c47c" (UID: "b520d752-90d4-41a1-a6d0-243e7058c47c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.958836 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-config\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.958954 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrthr\" (UniqueName: \"kubernetes.io/projected/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-kube-api-access-xrthr\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.959013 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-ovndb-tls-certs\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.959135 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-combined-ca-bundle\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.959187 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-httpd-config\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.959329 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54z7w\" (UniqueName: \"kubernetes.io/projected/b520d752-90d4-41a1-a6d0-243e7058c47c-kube-api-access-54z7w\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.959349 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.959358 4935 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.959366 4935 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.983862 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.986118 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-combined-ca-bundle\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.988212 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-config\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.989616 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-ovndb-tls-certs\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.991380 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrthr\" (UniqueName: \"kubernetes.io/projected/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-kube-api-access-xrthr\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.991502 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-config-data" (OuterVolumeSpecName: "config-data") pod "b520d752-90d4-41a1-a6d0-243e7058c47c" (UID: "b520d752-90d4-41a1-a6d0-243e7058c47c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:35 crc kubenswrapper[4935]: I1005 07:12:35.992047 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b520d752-90d4-41a1-a6d0-243e7058c47c" (UID: "b520d752-90d4-41a1-a6d0-243e7058c47c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.000532 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-httpd-config\") pod \"neutron-56b79f594d-q76hh\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.059458 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.061232 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.061252 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b520d752-90d4-41a1-a6d0-243e7058c47c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.174072 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.264754 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-swift-storage-0\") pod \"00decba6-51fa-4be0-b7be-e74327372758\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.264835 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-svc\") pod \"00decba6-51fa-4be0-b7be-e74327372758\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.264861 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-sb\") pod \"00decba6-51fa-4be0-b7be-e74327372758\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.264924 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-nb\") pod \"00decba6-51fa-4be0-b7be-e74327372758\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.264975 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5vn5\" (UniqueName: \"kubernetes.io/projected/00decba6-51fa-4be0-b7be-e74327372758-kube-api-access-j5vn5\") pod \"00decba6-51fa-4be0-b7be-e74327372758\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.264991 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-config\") pod \"00decba6-51fa-4be0-b7be-e74327372758\" (UID: \"00decba6-51fa-4be0-b7be-e74327372758\") " Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.277632 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6dbdfd7554-jrf6q"] Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.282207 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00decba6-51fa-4be0-b7be-e74327372758-kube-api-access-j5vn5" (OuterVolumeSpecName: "kube-api-access-j5vn5") pod "00decba6-51fa-4be0-b7be-e74327372758" (UID: "00decba6-51fa-4be0-b7be-e74327372758"). InnerVolumeSpecName "kube-api-access-j5vn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.294377 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.342435 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-config" (OuterVolumeSpecName: "config") pod "00decba6-51fa-4be0-b7be-e74327372758" (UID: "00decba6-51fa-4be0-b7be-e74327372758"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.343020 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00decba6-51fa-4be0-b7be-e74327372758" (UID: "00decba6-51fa-4be0-b7be-e74327372758"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.369024 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "00decba6-51fa-4be0-b7be-e74327372758" (UID: "00decba6-51fa-4be0-b7be-e74327372758"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.369406 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.369427 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.369438 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5vn5\" (UniqueName: \"kubernetes.io/projected/00decba6-51fa-4be0-b7be-e74327372758-kube-api-access-j5vn5\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.369452 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.370057 4935 generic.go:334] "Generic (PLEG): container finished" podID="00decba6-51fa-4be0-b7be-e74327372758" containerID="cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f" exitCode=0 Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.370144 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" event={"ID":"00decba6-51fa-4be0-b7be-e74327372758","Type":"ContainerDied","Data":"cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f"} Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.370174 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" event={"ID":"00decba6-51fa-4be0-b7be-e74327372758","Type":"ContainerDied","Data":"be1a5634d957122e1b48bea99487e888a3563a3c1a8d7f9d78e200947a232565"} Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.370192 4935 scope.go:117] "RemoveContainer" containerID="cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.370325 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6795bd975-gs8cj" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.376706 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00decba6-51fa-4be0-b7be-e74327372758" (UID: "00decba6-51fa-4be0-b7be-e74327372758"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.381827 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"980fa02a-09b0-42cc-8028-51b588427002","Type":"ContainerStarted","Data":"ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362"} Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.384162 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-k2ff8" event={"ID":"b520d752-90d4-41a1-a6d0-243e7058c47c","Type":"ContainerDied","Data":"38c577023ff6f5e54a282572ca4bb1643eb7324ba233f89f8f161a0a79a8a8c1"} Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.384186 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38c577023ff6f5e54a282572ca4bb1643eb7324ba233f89f8f161a0a79a8a8c1" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.384221 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-k2ff8" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.397841 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc97d9d55-72qqm"] Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.412246 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "00decba6-51fa-4be0-b7be-e74327372758" (UID: "00decba6-51fa-4be0-b7be-e74327372758"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.422935 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-79d9bc6587-tvhv2"] Oct 05 07:12:36 crc kubenswrapper[4935]: E1005 07:12:36.423526 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00decba6-51fa-4be0-b7be-e74327372758" containerName="dnsmasq-dns" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.423595 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="00decba6-51fa-4be0-b7be-e74327372758" containerName="dnsmasq-dns" Oct 05 07:12:36 crc kubenswrapper[4935]: E1005 07:12:36.423676 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00decba6-51fa-4be0-b7be-e74327372758" containerName="init" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.423724 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="00decba6-51fa-4be0-b7be-e74327372758" containerName="init" Oct 05 07:12:36 crc kubenswrapper[4935]: E1005 07:12:36.423777 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b520d752-90d4-41a1-a6d0-243e7058c47c" containerName="keystone-bootstrap" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.423826 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b520d752-90d4-41a1-a6d0-243e7058c47c" containerName="keystone-bootstrap" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.424042 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="00decba6-51fa-4be0-b7be-e74327372758" containerName="dnsmasq-dns" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.424114 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b520d752-90d4-41a1-a6d0-243e7058c47c" containerName="keystone-bootstrap" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.424713 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.426538 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.431474 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.431677 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xm842" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.432259 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.432352 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.432440 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.444400 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79d9bc6587-tvhv2"] Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.470777 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.470813 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00decba6-51fa-4be0-b7be-e74327372758-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.573275 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-combined-ca-bundle\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.573358 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-fernet-keys\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.573404 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-config-data\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.573441 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-credential-keys\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.573519 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-internal-tls-certs\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.573644 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-scripts\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.573714 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-public-tls-certs\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.573769 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbsd8\" (UniqueName: \"kubernetes.io/projected/6214d810-93b4-4537-a655-642df573048b-kube-api-access-sbsd8\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.674923 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-scripts\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.674985 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-public-tls-certs\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.675016 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbsd8\" (UniqueName: \"kubernetes.io/projected/6214d810-93b4-4537-a655-642df573048b-kube-api-access-sbsd8\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.675061 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-combined-ca-bundle\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.675082 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-fernet-keys\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.675103 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-config-data\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.675122 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-credential-keys\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.675159 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-internal-tls-certs\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.678293 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-scripts\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.678576 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-internal-tls-certs\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.681571 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-config-data\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.682254 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-fernet-keys\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.682300 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-public-tls-certs\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.682572 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-credential-keys\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.684476 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-combined-ca-bundle\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.702343 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbsd8\" (UniqueName: \"kubernetes.io/projected/6214d810-93b4-4537-a655-642df573048b-kube-api-access-sbsd8\") pod \"keystone-79d9bc6587-tvhv2\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.709211 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-gs8cj"] Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.718926 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-gs8cj"] Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.725454 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.743271 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.791077 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00decba6-51fa-4be0-b7be-e74327372758" path="/var/lib/kubelet/pods/00decba6-51fa-4be0-b7be-e74327372758/volumes" Oct 05 07:12:36 crc kubenswrapper[4935]: I1005 07:12:36.793555 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="058821c6-d6c2-442a-a5f2-652a60ab79f4" path="/var/lib/kubelet/pods/058821c6-d6c2-442a-a5f2-652a60ab79f4/volumes" Oct 05 07:12:37 crc kubenswrapper[4935]: W1005 07:12:37.879031 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2e01050_b5d1_496c_b24e_a2d1cdb4cb79.slice/crio-e03a97531abbe05ee0a2d911cef019568273b320014dfebbc03442aca7b4dac2 WatchSource:0}: Error finding container e03a97531abbe05ee0a2d911cef019568273b320014dfebbc03442aca7b4dac2: Status 404 returned error can't find the container with id e03a97531abbe05ee0a2d911cef019568273b320014dfebbc03442aca7b4dac2 Oct 05 07:12:37 crc kubenswrapper[4935]: W1005 07:12:37.880857 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88a7ceab_efbb_4ba4_a5f2_c6702abbc314.slice/crio-cf7c7ec1a50d259d4f557b17718ae497667bf183d921c9fe0b161a7bc0ba5a52 WatchSource:0}: Error finding container cf7c7ec1a50d259d4f557b17718ae497667bf183d921c9fe0b161a7bc0ba5a52: Status 404 returned error can't find the container with id cf7c7ec1a50d259d4f557b17718ae497667bf183d921c9fe0b161a7bc0ba5a52 Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.207056 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-56bb7fd5-dmzld"] Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.208448 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.212419 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.212723 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.254000 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56bb7fd5-dmzld"] Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.302939 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-public-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.303050 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7h4c\" (UniqueName: \"kubernetes.io/projected/27a28306-70be-4556-9659-64999f775195-kube-api-access-s7h4c\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.303074 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-ovndb-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.303096 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-httpd-config\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.303118 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-combined-ca-bundle\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.303134 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-config\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.303167 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-internal-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.404324 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7h4c\" (UniqueName: \"kubernetes.io/projected/27a28306-70be-4556-9659-64999f775195-kube-api-access-s7h4c\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.404363 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-ovndb-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.404400 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-httpd-config\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.404422 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-combined-ca-bundle\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.404439 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-config\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.404474 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-internal-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.404532 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-public-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.418851 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-ovndb-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.423881 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-public-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.426752 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-config\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.427532 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-combined-ca-bundle\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.428459 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-httpd-config\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.434810 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-internal-tls-certs\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.442444 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7h4c\" (UniqueName: \"kubernetes.io/projected/27a28306-70be-4556-9659-64999f775195-kube-api-access-s7h4c\") pod \"neutron-56bb7fd5-dmzld\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.455036 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" event={"ID":"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79","Type":"ContainerStarted","Data":"e03a97531abbe05ee0a2d911cef019568273b320014dfebbc03442aca7b4dac2"} Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.456658 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b885bd46-8ccc-47d6-8622-c4175e2267e5","Type":"ContainerStarted","Data":"7637814c072e95a7f542fd6a90d7331c8a2dfdb0f2b369847b4007abe680b762"} Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.457930 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbdfd7554-jrf6q" event={"ID":"88a7ceab-efbb-4ba4-a5f2-c6702abbc314","Type":"ContainerStarted","Data":"cf7c7ec1a50d259d4f557b17718ae497667bf183d921c9fe0b161a7bc0ba5a52"} Oct 05 07:12:38 crc kubenswrapper[4935]: I1005 07:12:38.532781 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:40 crc kubenswrapper[4935]: I1005 07:12:40.150738 4935 scope.go:117] "RemoveContainer" containerID="056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa" Oct 05 07:12:40 crc kubenswrapper[4935]: I1005 07:12:40.408081 4935 scope.go:117] "RemoveContainer" containerID="cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f" Oct 05 07:12:40 crc kubenswrapper[4935]: E1005 07:12:40.412113 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f\": container with ID starting with cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f not found: ID does not exist" containerID="cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f" Oct 05 07:12:40 crc kubenswrapper[4935]: I1005 07:12:40.412158 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f"} err="failed to get container status \"cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f\": rpc error: code = NotFound desc = could not find container \"cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f\": container with ID starting with cb26cf4d327fe77f27c8e528ba991af08fcb0174a69a5ca98e4dca21450f256f not found: ID does not exist" Oct 05 07:12:40 crc kubenswrapper[4935]: I1005 07:12:40.412186 4935 scope.go:117] "RemoveContainer" containerID="056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa" Oct 05 07:12:40 crc kubenswrapper[4935]: E1005 07:12:40.412764 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa\": container with ID starting with 056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa not found: ID does not exist" containerID="056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa" Oct 05 07:12:40 crc kubenswrapper[4935]: I1005 07:12:40.412797 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa"} err="failed to get container status \"056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa\": rpc error: code = NotFound desc = could not find container \"056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa\": container with ID starting with 056aaf62ad2694c4d40ae5f91576bdafbaab7dc14bec1ae7fed00ea0bc900caa not found: ID does not exist" Oct 05 07:12:40 crc kubenswrapper[4935]: I1005 07:12:40.637312 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79d9bc6587-tvhv2"] Oct 05 07:12:40 crc kubenswrapper[4935]: I1005 07:12:40.723103 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56b79f594d-q76hh"] Oct 05 07:12:40 crc kubenswrapper[4935]: W1005 07:12:40.741142 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeeb801c8_5f9c_4bfa_8f21_80a046fd04a8.slice/crio-0bd8ecbea4c2c8c483b0c8640819f1183b03355701c0b85985b269fe36e28c68 WatchSource:0}: Error finding container 0bd8ecbea4c2c8c483b0c8640819f1183b03355701c0b85985b269fe36e28c68: Status 404 returned error can't find the container with id 0bd8ecbea4c2c8c483b0c8640819f1183b03355701c0b85985b269fe36e28c68 Oct 05 07:12:40 crc kubenswrapper[4935]: I1005 07:12:40.833857 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56bb7fd5-dmzld"] Oct 05 07:12:40 crc kubenswrapper[4935]: W1005 07:12:40.844674 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27a28306_70be_4556_9659_64999f775195.slice/crio-cda4dff46d5e8b986aec637f976286eebb3d0f536a6e9ad90c6af7da74c71303 WatchSource:0}: Error finding container cda4dff46d5e8b986aec637f976286eebb3d0f536a6e9ad90c6af7da74c71303: Status 404 returned error can't find the container with id cda4dff46d5e8b986aec637f976286eebb3d0f536a6e9ad90c6af7da74c71303 Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.494137 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56b79f594d-q76hh" event={"ID":"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8","Type":"ContainerStarted","Data":"76684d743646083119a5adf79f28378bb03608f41db9fee37a119f5a9eba8108"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.494979 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56b79f594d-q76hh" event={"ID":"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8","Type":"ContainerStarted","Data":"ad39ce747967bfbf8195ccbec83da9a5e17c907f4a3c9ca8bd0259457636a718"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.495002 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.495015 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56b79f594d-q76hh" event={"ID":"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8","Type":"ContainerStarted","Data":"0bd8ecbea4c2c8c483b0c8640819f1183b03355701c0b85985b269fe36e28c68"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.497777 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b885bd46-8ccc-47d6-8622-c4175e2267e5","Type":"ContainerStarted","Data":"9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.499446 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56bb7fd5-dmzld" event={"ID":"27a28306-70be-4556-9659-64999f775195","Type":"ContainerStarted","Data":"a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.499470 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56bb7fd5-dmzld" event={"ID":"27a28306-70be-4556-9659-64999f775195","Type":"ContainerStarted","Data":"cda4dff46d5e8b986aec637f976286eebb3d0f536a6e9ad90c6af7da74c71303"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.501122 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbdfd7554-jrf6q" event={"ID":"88a7ceab-efbb-4ba4-a5f2-c6702abbc314","Type":"ContainerStarted","Data":"829086067b4c67ac2011dcae6d0f8b4b1d850acfb49e4cf8b466a30f81bb3be2"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.501150 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbdfd7554-jrf6q" event={"ID":"88a7ceab-efbb-4ba4-a5f2-c6702abbc314","Type":"ContainerStarted","Data":"27997ed35ed34c0a762315ae80ed4c65e006d4cd919243a41130b860e5a4e10a"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.501215 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.501261 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.503466 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerStarted","Data":"b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.505522 4935 generic.go:334] "Generic (PLEG): container finished" podID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerID="51b2ef43d64a48354781cfd1e67ed1fa7c882a3527b2e0cffe22d77f0a02679f" exitCode=0 Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.505573 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" event={"ID":"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79","Type":"ContainerDied","Data":"51b2ef43d64a48354781cfd1e67ed1fa7c882a3527b2e0cffe22d77f0a02679f"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.507496 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"980fa02a-09b0-42cc-8028-51b588427002","Type":"ContainerStarted","Data":"9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.510015 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79d9bc6587-tvhv2" event={"ID":"6214d810-93b4-4537-a655-642df573048b","Type":"ContainerStarted","Data":"010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.510070 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79d9bc6587-tvhv2" event={"ID":"6214d810-93b4-4537-a655-642df573048b","Type":"ContainerStarted","Data":"2ddd8c31b1bc7ebcb2278e234b4f7aa4d87109bcbd539f6f21dc2ef468d8bc2e"} Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.510257 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.526723 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-56b79f594d-q76hh" podStartSLOduration=6.526704885 podStartE2EDuration="6.526704885s" podCreationTimestamp="2025-10-05 07:12:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:41.518016684 +0000 UTC m=+1195.400643164" watchObservedRunningTime="2025-10-05 07:12:41.526704885 +0000 UTC m=+1195.409331355" Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.573653 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-79d9bc6587-tvhv2" podStartSLOduration=5.573632783 podStartE2EDuration="5.573632783s" podCreationTimestamp="2025-10-05 07:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:41.547253811 +0000 UTC m=+1195.429880261" watchObservedRunningTime="2025-10-05 07:12:41.573632783 +0000 UTC m=+1195.456259263" Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.614140 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6dbdfd7554-jrf6q" podStartSLOduration=6.614122711 podStartE2EDuration="6.614122711s" podCreationTimestamp="2025-10-05 07:12:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:41.593377659 +0000 UTC m=+1195.476004119" watchObservedRunningTime="2025-10-05 07:12:41.614122711 +0000 UTC m=+1195.496749171" Oct 05 07:12:41 crc kubenswrapper[4935]: I1005 07:12:41.631425 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.631408671 podStartE2EDuration="7.631408671s" podCreationTimestamp="2025-10-05 07:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:41.622122054 +0000 UTC m=+1195.504748514" watchObservedRunningTime="2025-10-05 07:12:41.631408671 +0000 UTC m=+1195.514035131" Oct 05 07:12:42 crc kubenswrapper[4935]: I1005 07:12:42.524570 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56bb7fd5-dmzld" event={"ID":"27a28306-70be-4556-9659-64999f775195","Type":"ContainerStarted","Data":"16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000"} Oct 05 07:12:42 crc kubenswrapper[4935]: I1005 07:12:42.525386 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:12:42 crc kubenswrapper[4935]: I1005 07:12:42.538806 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" event={"ID":"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79","Type":"ContainerStarted","Data":"2f9de2f2ec63a3b30bf4e29252e9801a9c005e253ce47dd65f96d3a34902f706"} Oct 05 07:12:42 crc kubenswrapper[4935]: I1005 07:12:42.538941 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:42 crc kubenswrapper[4935]: I1005 07:12:42.542345 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b885bd46-8ccc-47d6-8622-c4175e2267e5","Type":"ContainerStarted","Data":"9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c"} Oct 05 07:12:42 crc kubenswrapper[4935]: I1005 07:12:42.557722 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-56bb7fd5-dmzld" podStartSLOduration=4.557697295 podStartE2EDuration="4.557697295s" podCreationTimestamp="2025-10-05 07:12:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:42.54585616 +0000 UTC m=+1196.428482630" watchObservedRunningTime="2025-10-05 07:12:42.557697295 +0000 UTC m=+1196.440323755" Oct 05 07:12:42 crc kubenswrapper[4935]: I1005 07:12:42.586397 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.586372508 podStartE2EDuration="7.586372508s" podCreationTimestamp="2025-10-05 07:12:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:42.577953504 +0000 UTC m=+1196.460580004" watchObservedRunningTime="2025-10-05 07:12:42.586372508 +0000 UTC m=+1196.468998958" Oct 05 07:12:42 crc kubenswrapper[4935]: I1005 07:12:42.606670 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" podStartSLOduration=7.606642287 podStartE2EDuration="7.606642287s" podCreationTimestamp="2025-10-05 07:12:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:42.604505921 +0000 UTC m=+1196.487132421" watchObservedRunningTime="2025-10-05 07:12:42.606642287 +0000 UTC m=+1196.489268757" Oct 05 07:12:44 crc kubenswrapper[4935]: I1005 07:12:44.661611 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:44 crc kubenswrapper[4935]: I1005 07:12:44.661945 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:44 crc kubenswrapper[4935]: I1005 07:12:44.707746 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:44 crc kubenswrapper[4935]: I1005 07:12:44.712352 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:45 crc kubenswrapper[4935]: I1005 07:12:45.573522 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n9vr5" event={"ID":"e9ba8cbb-45b9-49e5-b5a0-9277b724b607","Type":"ContainerStarted","Data":"f72801a860065d79604fdd54e0db50ff9192f050ef2dd2fefaa717d191b2c075"} Oct 05 07:12:45 crc kubenswrapper[4935]: I1005 07:12:45.573940 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:45 crc kubenswrapper[4935]: I1005 07:12:45.573980 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:45 crc kubenswrapper[4935]: I1005 07:12:45.598689 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-n9vr5" podStartSLOduration=3.013694427 podStartE2EDuration="36.598668405s" podCreationTimestamp="2025-10-05 07:12:09 +0000 UTC" firstStartedPulling="2025-10-05 07:12:10.657169655 +0000 UTC m=+1164.539796115" lastFinishedPulling="2025-10-05 07:12:44.242143633 +0000 UTC m=+1198.124770093" observedRunningTime="2025-10-05 07:12:45.594994187 +0000 UTC m=+1199.477620647" watchObservedRunningTime="2025-10-05 07:12:45.598668405 +0000 UTC m=+1199.481294865" Oct 05 07:12:46 crc kubenswrapper[4935]: I1005 07:12:46.060802 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 07:12:46 crc kubenswrapper[4935]: I1005 07:12:46.061083 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 07:12:46 crc kubenswrapper[4935]: I1005 07:12:46.088861 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 07:12:46 crc kubenswrapper[4935]: I1005 07:12:46.098643 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 07:12:46 crc kubenswrapper[4935]: I1005 07:12:46.586139 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 07:12:46 crc kubenswrapper[4935]: I1005 07:12:46.586175 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 07:12:47 crc kubenswrapper[4935]: I1005 07:12:47.601147 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:47 crc kubenswrapper[4935]: I1005 07:12:47.601513 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 07:12:47 crc kubenswrapper[4935]: I1005 07:12:47.605505 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 07:12:48 crc kubenswrapper[4935]: I1005 07:12:48.559767 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 07:12:48 crc kubenswrapper[4935]: I1005 07:12:48.562698 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.197015 4935 scope.go:117] "RemoveContainer" containerID="dfe62c90fa335a0ed334f9839281e87b352861197b736d726472fd5a4328c441" Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.226317 4935 scope.go:117] "RemoveContainer" containerID="8de0256dd007a5a553144b46aad348d4313786ac0936ce0e5e7a3bf1b3212b3c" Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.252256 4935 scope.go:117] "RemoveContainer" containerID="1b99a395eca4a926f0911f71ad49e3c7b4f9a7ca96d1fa3f05b7b1557b0ef1f9" Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.628809 4935 generic.go:334] "Generic (PLEG): container finished" podID="e9ba8cbb-45b9-49e5-b5a0-9277b724b607" containerID="f72801a860065d79604fdd54e0db50ff9192f050ef2dd2fefaa717d191b2c075" exitCode=0 Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.628914 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n9vr5" event={"ID":"e9ba8cbb-45b9-49e5-b5a0-9277b724b607","Type":"ContainerDied","Data":"f72801a860065d79604fdd54e0db50ff9192f050ef2dd2fefaa717d191b2c075"} Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.631300 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nl9vd" event={"ID":"2ef780aa-5f7e-4fa2-a019-6ee2c3993692","Type":"ContainerStarted","Data":"068f0258b4f42367cfe09495e247e8ae26cd3ea724102859b0fd3528a91385e1"} Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.634461 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerStarted","Data":"7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65"} Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.634715 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.634674 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="ceilometer-central-agent" containerID="cri-o://7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde" gracePeriod=30 Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.634770 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="sg-core" containerID="cri-o://b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d" gracePeriod=30 Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.634815 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="ceilometer-notification-agent" containerID="cri-o://52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06" gracePeriod=30 Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.634839 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="proxy-httpd" containerID="cri-o://7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65" gracePeriod=30 Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.686128 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.499399629 podStartE2EDuration="41.686099035s" podCreationTimestamp="2025-10-05 07:12:09 +0000 UTC" firstStartedPulling="2025-10-05 07:12:10.239513147 +0000 UTC m=+1164.122139607" lastFinishedPulling="2025-10-05 07:12:49.426212543 +0000 UTC m=+1203.308839013" observedRunningTime="2025-10-05 07:12:50.674459475 +0000 UTC m=+1204.557086015" watchObservedRunningTime="2025-10-05 07:12:50.686099035 +0000 UTC m=+1204.568725535" Oct 05 07:12:50 crc kubenswrapper[4935]: I1005 07:12:50.985034 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.014608 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-nl9vd" podStartSLOduration=3.241072049 podStartE2EDuration="42.014574404s" podCreationTimestamp="2025-10-05 07:12:09 +0000 UTC" firstStartedPulling="2025-10-05 07:12:10.64076743 +0000 UTC m=+1164.523393890" lastFinishedPulling="2025-10-05 07:12:49.414269775 +0000 UTC m=+1203.296896245" observedRunningTime="2025-10-05 07:12:50.712684892 +0000 UTC m=+1204.595311362" watchObservedRunningTime="2025-10-05 07:12:51.014574404 +0000 UTC m=+1204.897200874" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.049607 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-cw86l"] Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.049846 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77594f5649-cw86l" podUID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" containerName="dnsmasq-dns" containerID="cri-o://4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e" gracePeriod=10 Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.566797 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.646526 4935 generic.go:334] "Generic (PLEG): container finished" podID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" containerID="4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e" exitCode=0 Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.646590 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77594f5649-cw86l" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.646611 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-cw86l" event={"ID":"333447cb-9bb5-4e0f-bbd0-c320cecc108d","Type":"ContainerDied","Data":"4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e"} Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.646671 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-cw86l" event={"ID":"333447cb-9bb5-4e0f-bbd0-c320cecc108d","Type":"ContainerDied","Data":"d7a3841b39e2daded300aa80e8ef8e775b40e6c82db9f04d5d0c94f98a09b708"} Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.646701 4935 scope.go:117] "RemoveContainer" containerID="4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.649818 4935 generic.go:334] "Generic (PLEG): container finished" podID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerID="7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65" exitCode=0 Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.649847 4935 generic.go:334] "Generic (PLEG): container finished" podID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerID="b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d" exitCode=2 Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.649856 4935 generic.go:334] "Generic (PLEG): container finished" podID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerID="7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde" exitCode=0 Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.649966 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerDied","Data":"7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65"} Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.650033 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerDied","Data":"b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d"} Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.650056 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerDied","Data":"7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde"} Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.682107 4935 scope.go:117] "RemoveContainer" containerID="c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.692871 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-swift-storage-0\") pod \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.692955 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-svc\") pod \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.692992 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-sb\") pod \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.693016 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt589\" (UniqueName: \"kubernetes.io/projected/333447cb-9bb5-4e0f-bbd0-c320cecc108d-kube-api-access-vt589\") pod \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.693137 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-config\") pod \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.694205 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-nb\") pod \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\" (UID: \"333447cb-9bb5-4e0f-bbd0-c320cecc108d\") " Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.700468 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333447cb-9bb5-4e0f-bbd0-c320cecc108d-kube-api-access-vt589" (OuterVolumeSpecName: "kube-api-access-vt589") pod "333447cb-9bb5-4e0f-bbd0-c320cecc108d" (UID: "333447cb-9bb5-4e0f-bbd0-c320cecc108d"). InnerVolumeSpecName "kube-api-access-vt589". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.711390 4935 scope.go:117] "RemoveContainer" containerID="4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e" Oct 05 07:12:51 crc kubenswrapper[4935]: E1005 07:12:51.716120 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e\": container with ID starting with 4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e not found: ID does not exist" containerID="4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.716172 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e"} err="failed to get container status \"4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e\": rpc error: code = NotFound desc = could not find container \"4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e\": container with ID starting with 4430283d4709f090bbaecbed9bc7407bbd81ec731084f740cc14e408fb2e208e not found: ID does not exist" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.716201 4935 scope.go:117] "RemoveContainer" containerID="c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38" Oct 05 07:12:51 crc kubenswrapper[4935]: E1005 07:12:51.717504 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38\": container with ID starting with c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38 not found: ID does not exist" containerID="c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.717542 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38"} err="failed to get container status \"c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38\": rpc error: code = NotFound desc = could not find container \"c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38\": container with ID starting with c7f1c5f94a14078db76bb8cb41f8882e11a1d61a2fcb7c22096f0e166e128c38 not found: ID does not exist" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.757451 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "333447cb-9bb5-4e0f-bbd0-c320cecc108d" (UID: "333447cb-9bb5-4e0f-bbd0-c320cecc108d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.765513 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-config" (OuterVolumeSpecName: "config") pod "333447cb-9bb5-4e0f-bbd0-c320cecc108d" (UID: "333447cb-9bb5-4e0f-bbd0-c320cecc108d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.786172 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "333447cb-9bb5-4e0f-bbd0-c320cecc108d" (UID: "333447cb-9bb5-4e0f-bbd0-c320cecc108d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.793973 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "333447cb-9bb5-4e0f-bbd0-c320cecc108d" (UID: "333447cb-9bb5-4e0f-bbd0-c320cecc108d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.796302 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.796324 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.796334 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.796342 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.796350 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt589\" (UniqueName: \"kubernetes.io/projected/333447cb-9bb5-4e0f-bbd0-c320cecc108d-kube-api-access-vt589\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.801118 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "333447cb-9bb5-4e0f-bbd0-c320cecc108d" (UID: "333447cb-9bb5-4e0f-bbd0-c320cecc108d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:51 crc kubenswrapper[4935]: I1005 07:12:51.898280 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333447cb-9bb5-4e0f-bbd0-c320cecc108d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.019507 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.034631 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-cw86l"] Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.044476 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-cw86l"] Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.101552 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-config-data\") pod \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.101628 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-combined-ca-bundle\") pod \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.101673 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-scripts\") pod \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.101757 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-db-sync-config-data\") pod \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.101775 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z26l2\" (UniqueName: \"kubernetes.io/projected/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-kube-api-access-z26l2\") pod \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.101827 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-etc-machine-id\") pod \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\" (UID: \"e9ba8cbb-45b9-49e5-b5a0-9277b724b607\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.102201 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e9ba8cbb-45b9-49e5-b5a0-9277b724b607" (UID: "e9ba8cbb-45b9-49e5-b5a0-9277b724b607"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.107089 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-kube-api-access-z26l2" (OuterVolumeSpecName: "kube-api-access-z26l2") pod "e9ba8cbb-45b9-49e5-b5a0-9277b724b607" (UID: "e9ba8cbb-45b9-49e5-b5a0-9277b724b607"). InnerVolumeSpecName "kube-api-access-z26l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.107162 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e9ba8cbb-45b9-49e5-b5a0-9277b724b607" (UID: "e9ba8cbb-45b9-49e5-b5a0-9277b724b607"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.107328 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-scripts" (OuterVolumeSpecName: "scripts") pod "e9ba8cbb-45b9-49e5-b5a0-9277b724b607" (UID: "e9ba8cbb-45b9-49e5-b5a0-9277b724b607"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.129273 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9ba8cbb-45b9-49e5-b5a0-9277b724b607" (UID: "e9ba8cbb-45b9-49e5-b5a0-9277b724b607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.168597 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-config-data" (OuterVolumeSpecName: "config-data") pod "e9ba8cbb-45b9-49e5-b5a0-9277b724b607" (UID: "e9ba8cbb-45b9-49e5-b5a0-9277b724b607"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.203662 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.203693 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.203706 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.203714 4935 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.203722 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z26l2\" (UniqueName: \"kubernetes.io/projected/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-kube-api-access-z26l2\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.203731 4935 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9ba8cbb-45b9-49e5-b5a0-9277b724b607-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.235882 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.406368 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-combined-ca-bundle\") pod \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.406717 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2f2s\" (UniqueName: \"kubernetes.io/projected/8c71f889-f8cd-4f42-a125-faaaaa1d3501-kube-api-access-s2f2s\") pod \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.406749 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-scripts\") pod \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.406796 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-sg-core-conf-yaml\") pod \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.406852 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-config-data\") pod \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.406914 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-run-httpd\") pod \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.406952 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-log-httpd\") pod \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\" (UID: \"8c71f889-f8cd-4f42-a125-faaaaa1d3501\") " Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.407734 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8c71f889-f8cd-4f42-a125-faaaaa1d3501" (UID: "8c71f889-f8cd-4f42-a125-faaaaa1d3501"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.408058 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8c71f889-f8cd-4f42-a125-faaaaa1d3501" (UID: "8c71f889-f8cd-4f42-a125-faaaaa1d3501"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.410800 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c71f889-f8cd-4f42-a125-faaaaa1d3501-kube-api-access-s2f2s" (OuterVolumeSpecName: "kube-api-access-s2f2s") pod "8c71f889-f8cd-4f42-a125-faaaaa1d3501" (UID: "8c71f889-f8cd-4f42-a125-faaaaa1d3501"). InnerVolumeSpecName "kube-api-access-s2f2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.412178 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-scripts" (OuterVolumeSpecName: "scripts") pod "8c71f889-f8cd-4f42-a125-faaaaa1d3501" (UID: "8c71f889-f8cd-4f42-a125-faaaaa1d3501"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.432790 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8c71f889-f8cd-4f42-a125-faaaaa1d3501" (UID: "8c71f889-f8cd-4f42-a125-faaaaa1d3501"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.501360 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-config-data" (OuterVolumeSpecName: "config-data") pod "8c71f889-f8cd-4f42-a125-faaaaa1d3501" (UID: "8c71f889-f8cd-4f42-a125-faaaaa1d3501"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.510175 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2f2s\" (UniqueName: \"kubernetes.io/projected/8c71f889-f8cd-4f42-a125-faaaaa1d3501-kube-api-access-s2f2s\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.510230 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.510253 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.510274 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.510293 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.510315 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c71f889-f8cd-4f42-a125-faaaaa1d3501-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.512622 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c71f889-f8cd-4f42-a125-faaaaa1d3501" (UID: "8c71f889-f8cd-4f42-a125-faaaaa1d3501"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.611644 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c71f889-f8cd-4f42-a125-faaaaa1d3501-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.690669 4935 generic.go:334] "Generic (PLEG): container finished" podID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerID="52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06" exitCode=0 Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.690740 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerDied","Data":"52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06"} Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.690774 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c71f889-f8cd-4f42-a125-faaaaa1d3501","Type":"ContainerDied","Data":"e47551e80dc05743699b4d5fa2006d5e7f6fd8a591227c880cc7668f115e2e0c"} Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.690797 4935 scope.go:117] "RemoveContainer" containerID="7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.690955 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.697970 4935 generic.go:334] "Generic (PLEG): container finished" podID="2ef780aa-5f7e-4fa2-a019-6ee2c3993692" containerID="068f0258b4f42367cfe09495e247e8ae26cd3ea724102859b0fd3528a91385e1" exitCode=0 Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.698025 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nl9vd" event={"ID":"2ef780aa-5f7e-4fa2-a019-6ee2c3993692","Type":"ContainerDied","Data":"068f0258b4f42367cfe09495e247e8ae26cd3ea724102859b0fd3528a91385e1"} Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.704611 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n9vr5" event={"ID":"e9ba8cbb-45b9-49e5-b5a0-9277b724b607","Type":"ContainerDied","Data":"54666955fce9b36f2a70f3c5c88e9a86a09098ad74e7e05f6db9dcb24c298d53"} Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.704658 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54666955fce9b36f2a70f3c5c88e9a86a09098ad74e7e05f6db9dcb24c298d53" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.704730 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n9vr5" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.743650 4935 scope.go:117] "RemoveContainer" containerID="b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.746601 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.769236 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.800675 4935 scope.go:117] "RemoveContainer" containerID="52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.828684 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" path="/var/lib/kubelet/pods/333447cb-9bb5-4e0f-bbd0-c320cecc108d/volumes" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.829700 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" path="/var/lib/kubelet/pods/8c71f889-f8cd-4f42-a125-faaaaa1d3501/volumes" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.830467 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.830799 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="ceilometer-notification-agent" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.830859 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="ceilometer-notification-agent" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.830978 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="sg-core" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.831054 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="sg-core" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.831131 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="ceilometer-central-agent" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.831188 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="ceilometer-central-agent" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.831262 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ba8cbb-45b9-49e5-b5a0-9277b724b607" containerName="cinder-db-sync" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.831310 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ba8cbb-45b9-49e5-b5a0-9277b724b607" containerName="cinder-db-sync" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.831372 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" containerName="init" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.831434 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" containerName="init" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.831496 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" containerName="dnsmasq-dns" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.831550 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" containerName="dnsmasq-dns" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.831608 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="proxy-httpd" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.831668 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="proxy-httpd" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.836497 4935 scope.go:117] "RemoveContainer" containerID="7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.837417 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="proxy-httpd" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.837490 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="333447cb-9bb5-4e0f-bbd0-c320cecc108d" containerName="dnsmasq-dns" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.837558 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ba8cbb-45b9-49e5-b5a0-9277b724b607" containerName="cinder-db-sync" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.837631 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="ceilometer-central-agent" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.837690 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="ceilometer-notification-agent" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.837760 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c71f889-f8cd-4f42-a125-faaaaa1d3501" containerName="sg-core" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.840704 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.840790 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.843352 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.843683 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.857693 4935 scope.go:117] "RemoveContainer" containerID="7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.859003 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65\": container with ID starting with 7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65 not found: ID does not exist" containerID="7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.859037 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65"} err="failed to get container status \"7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65\": rpc error: code = NotFound desc = could not find container \"7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65\": container with ID starting with 7e2169279c1574758a76c197487a2da6808cededfb930bba4ef661c319f19f65 not found: ID does not exist" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.859066 4935 scope.go:117] "RemoveContainer" containerID="b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.862481 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d\": container with ID starting with b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d not found: ID does not exist" containerID="b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.862522 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d"} err="failed to get container status \"b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d\": rpc error: code = NotFound desc = could not find container \"b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d\": container with ID starting with b0267c24c4fdcbd46c93ea9e959d7cdffeb038283a3cd9c78b658616f6e5c31d not found: ID does not exist" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.862546 4935 scope.go:117] "RemoveContainer" containerID="52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.862957 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06\": container with ID starting with 52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06 not found: ID does not exist" containerID="52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.863008 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06"} err="failed to get container status \"52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06\": rpc error: code = NotFound desc = could not find container \"52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06\": container with ID starting with 52d2a03843c54e95252ce71eaa1d9086c93b7772d67a383e7effc900a2efbf06 not found: ID does not exist" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.863043 4935 scope.go:117] "RemoveContainer" containerID="7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde" Oct 05 07:12:52 crc kubenswrapper[4935]: E1005 07:12:52.863568 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde\": container with ID starting with 7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde not found: ID does not exist" containerID="7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.863609 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde"} err="failed to get container status \"7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde\": rpc error: code = NotFound desc = could not find container \"7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde\": container with ID starting with 7d0356c44e82131981ee8398abed48994a6f59b496f2c65564a51608d116edde not found: ID does not exist" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.947241 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.948904 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.952121 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.952372 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.952537 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.952552 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-tpbjr" Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.961839 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.990881 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77b76d77dc-clsjp"] Oct 05 07:12:52 crc kubenswrapper[4935]: I1005 07:12:52.992399 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.014089 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77b76d77dc-clsjp"] Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.029296 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-config-data\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.029341 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.029390 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-scripts\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.029506 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-log-httpd\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.029537 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.029555 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv9ll\" (UniqueName: \"kubernetes.io/projected/6b527f48-ab8c-41e2-89b6-08e2456c687a-kube-api-access-cv9ll\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.029578 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-run-httpd\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131122 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-scripts\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131179 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x68v4\" (UniqueName: \"kubernetes.io/projected/20540188-e24e-48b4-acde-2214c8b4d2b9-kube-api-access-x68v4\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131211 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131230 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-sb\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131253 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-config-data\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131302 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131329 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-config\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131346 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-svc\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131365 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131381 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-swift-storage-0\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131591 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-scripts\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131622 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f66f223c-e87c-4314-957f-bd7c55dafc8f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131645 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8cpx\" (UniqueName: \"kubernetes.io/projected/f66f223c-e87c-4314-957f-bd7c55dafc8f-kube-api-access-v8cpx\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131681 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131703 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-nb\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.131993 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-log-httpd\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.132092 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.132146 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv9ll\" (UniqueName: \"kubernetes.io/projected/6b527f48-ab8c-41e2-89b6-08e2456c687a-kube-api-access-cv9ll\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.132220 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-run-httpd\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.132707 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-run-httpd\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.132922 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-log-httpd\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.139143 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-scripts\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.140122 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-config-data\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.140422 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.143488 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.148518 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.149987 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.160048 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.162475 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv9ll\" (UniqueName: \"kubernetes.io/projected/6b527f48-ab8c-41e2-89b6-08e2456c687a-kube-api-access-cv9ll\") pod \"ceilometer-0\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.189653 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.234768 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235053 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-nb\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235124 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-scripts\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235147 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x68v4\" (UniqueName: \"kubernetes.io/projected/20540188-e24e-48b4-acde-2214c8b4d2b9-kube-api-access-x68v4\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235170 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235189 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-sb\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235214 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-config\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235236 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-svc\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235259 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235279 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-swift-storage-0\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235320 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f66f223c-e87c-4314-957f-bd7c55dafc8f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.235343 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8cpx\" (UniqueName: \"kubernetes.io/projected/f66f223c-e87c-4314-957f-bd7c55dafc8f-kube-api-access-v8cpx\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.237572 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-sb\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.240630 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-nb\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.241187 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-config\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.241678 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-svc\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.242219 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-swift-storage-0\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.242259 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f66f223c-e87c-4314-957f-bd7c55dafc8f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.255501 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.255858 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.256307 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-scripts\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.257037 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x68v4\" (UniqueName: \"kubernetes.io/projected/20540188-e24e-48b4-acde-2214c8b4d2b9-kube-api-access-x68v4\") pod \"dnsmasq-dns-77b76d77dc-clsjp\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.260589 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8cpx\" (UniqueName: \"kubernetes.io/projected/f66f223c-e87c-4314-957f-bd7c55dafc8f-kube-api-access-v8cpx\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.260909 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.270280 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.312906 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.338876 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.338945 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.338964 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7tnl\" (UniqueName: \"kubernetes.io/projected/e94e21e9-e1a5-406b-9235-35165cded309-kube-api-access-w7tnl\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.338999 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-scripts\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.339088 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e94e21e9-e1a5-406b-9235-35165cded309-logs\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.339192 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data-custom\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.339210 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e94e21e9-e1a5-406b-9235-35165cded309-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.441233 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e94e21e9-e1a5-406b-9235-35165cded309-logs\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.441281 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data-custom\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.441306 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e94e21e9-e1a5-406b-9235-35165cded309-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.441405 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.441464 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.441486 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7tnl\" (UniqueName: \"kubernetes.io/projected/e94e21e9-e1a5-406b-9235-35165cded309-kube-api-access-w7tnl\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.441561 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-scripts\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.442044 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e94e21e9-e1a5-406b-9235-35165cded309-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.442461 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e94e21e9-e1a5-406b-9235-35165cded309-logs\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.450511 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.452337 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.460311 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.463500 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-scripts\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.463847 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7tnl\" (UniqueName: \"kubernetes.io/projected/e94e21e9-e1a5-406b-9235-35165cded309-kube-api-access-w7tnl\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.482663 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data-custom\") pod \"cinder-api-0\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.613043 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.752845 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.831802 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77b76d77dc-clsjp"] Oct 05 07:12:53 crc kubenswrapper[4935]: W1005 07:12:53.839439 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20540188_e24e_48b4_acde_2214c8b4d2b9.slice/crio-d249b1162e1a38677438f864e5f05a09ca0f14734edc43f4840987d391057dfd WatchSource:0}: Error finding container d249b1162e1a38677438f864e5f05a09ca0f14734edc43f4840987d391057dfd: Status 404 returned error can't find the container with id d249b1162e1a38677438f864e5f05a09ca0f14734edc43f4840987d391057dfd Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.933105 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:12:53 crc kubenswrapper[4935]: I1005 07:12:53.971898 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:12:53 crc kubenswrapper[4935]: W1005 07:12:53.980427 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b527f48_ab8c_41e2_89b6_08e2456c687a.slice/crio-21bc9df07348db72d14c7b356af2f4f90e9f138012469e9f0bfc480eecb2253a WatchSource:0}: Error finding container 21bc9df07348db72d14c7b356af2f4f90e9f138012469e9f0bfc480eecb2253a: Status 404 returned error can't find the container with id 21bc9df07348db72d14c7b356af2f4f90e9f138012469e9f0bfc480eecb2253a Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.072061 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.256251 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-db-sync-config-data\") pod \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.256397 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-combined-ca-bundle\") pod \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.256458 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skwwn\" (UniqueName: \"kubernetes.io/projected/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-kube-api-access-skwwn\") pod \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\" (UID: \"2ef780aa-5f7e-4fa2-a019-6ee2c3993692\") " Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.261983 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2ef780aa-5f7e-4fa2-a019-6ee2c3993692" (UID: "2ef780aa-5f7e-4fa2-a019-6ee2c3993692"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.261997 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-kube-api-access-skwwn" (OuterVolumeSpecName: "kube-api-access-skwwn") pod "2ef780aa-5f7e-4fa2-a019-6ee2c3993692" (UID: "2ef780aa-5f7e-4fa2-a019-6ee2c3993692"). InnerVolumeSpecName "kube-api-access-skwwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.290271 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ef780aa-5f7e-4fa2-a019-6ee2c3993692" (UID: "2ef780aa-5f7e-4fa2-a019-6ee2c3993692"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.358375 4935 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.358628 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.358795 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skwwn\" (UniqueName: \"kubernetes.io/projected/2ef780aa-5f7e-4fa2-a019-6ee2c3993692-kube-api-access-skwwn\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.771550 4935 generic.go:334] "Generic (PLEG): container finished" podID="20540188-e24e-48b4-acde-2214c8b4d2b9" containerID="ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3" exitCode=0 Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.772674 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" event={"ID":"20540188-e24e-48b4-acde-2214c8b4d2b9","Type":"ContainerDied","Data":"ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3"} Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.772730 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" event={"ID":"20540188-e24e-48b4-acde-2214c8b4d2b9","Type":"ContainerStarted","Data":"d249b1162e1a38677438f864e5f05a09ca0f14734edc43f4840987d391057dfd"} Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.783883 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nl9vd" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.808171 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e94e21e9-e1a5-406b-9235-35165cded309","Type":"ContainerStarted","Data":"e57b27b19e3967d9be9e7a1b859310794e3bc546a933ee59a48b2cc0ef49ea92"} Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.808220 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e94e21e9-e1a5-406b-9235-35165cded309","Type":"ContainerStarted","Data":"f47cbeb7226c7fb1e8f46f5b604659173cebbbde55bea4e77c469cd787378f1a"} Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.808236 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerStarted","Data":"21bc9df07348db72d14c7b356af2f4f90e9f138012469e9f0bfc480eecb2253a"} Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.808248 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f66f223c-e87c-4314-957f-bd7c55dafc8f","Type":"ContainerStarted","Data":"546d24e9310c8d9cd95ff13ba5c5d2410e8fd137dfabbe554bb38aa898d2367d"} Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.808262 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nl9vd" event={"ID":"2ef780aa-5f7e-4fa2-a019-6ee2c3993692","Type":"ContainerDied","Data":"7227578270f1512894036edc79f46ad87e6b2db92de39cacf28ea2b2be313fb5"} Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.808276 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7227578270f1512894036edc79f46ad87e6b2db92de39cacf28ea2b2be313fb5" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.978837 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7cbc6494d6-8n275"] Oct 05 07:12:54 crc kubenswrapper[4935]: E1005 07:12:54.979783 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef780aa-5f7e-4fa2-a019-6ee2c3993692" containerName="barbican-db-sync" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.979797 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef780aa-5f7e-4fa2-a019-6ee2c3993692" containerName="barbican-db-sync" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.979983 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef780aa-5f7e-4fa2-a019-6ee2c3993692" containerName="barbican-db-sync" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.980996 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.983691 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.983967 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x7t8h" Oct 05 07:12:54 crc kubenswrapper[4935]: I1005 07:12:54.984467 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.000948 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-57fd77d677-fr4zj"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.011351 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.013368 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.024029 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7cbc6494d6-8n275"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.041194 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57fd77d677-fr4zj"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073127 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-combined-ca-bundle\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073205 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073239 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-combined-ca-bundle\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073263 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq8v4\" (UniqueName: \"kubernetes.io/projected/c9b2e8af-97b0-4103-a241-ca473f299043-kube-api-access-lq8v4\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073286 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073321 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9b2e8af-97b0-4103-a241-ca473f299043-logs\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073347 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data-custom\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073372 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data-custom\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073391 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-logs\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.073411 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkx65\" (UniqueName: \"kubernetes.io/projected/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-kube-api-access-qkx65\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.106758 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77b76d77dc-clsjp"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.145200 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-k9hln"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.157111 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.169944 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-k9hln"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174597 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-combined-ca-bundle\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174665 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174697 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-combined-ca-bundle\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174720 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq8v4\" (UniqueName: \"kubernetes.io/projected/c9b2e8af-97b0-4103-a241-ca473f299043-kube-api-access-lq8v4\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174743 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174775 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9b2e8af-97b0-4103-a241-ca473f299043-logs\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174799 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data-custom\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174826 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data-custom\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174861 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-logs\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.174881 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkx65\" (UniqueName: \"kubernetes.io/projected/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-kube-api-access-qkx65\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.176140 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9b2e8af-97b0-4103-a241-ca473f299043-logs\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.177310 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-logs\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.180667 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-75db496546-lvnqq"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.184439 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.184876 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.185561 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-combined-ca-bundle\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.187571 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data-custom\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.188175 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.188799 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.190150 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-combined-ca-bundle\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.191274 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75db496546-lvnqq"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.197936 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data-custom\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.202120 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.205391 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq8v4\" (UniqueName: \"kubernetes.io/projected/c9b2e8af-97b0-4103-a241-ca473f299043-kube-api-access-lq8v4\") pod \"barbican-worker-57fd77d677-fr4zj\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.205949 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkx65\" (UniqueName: \"kubernetes.io/projected/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-kube-api-access-qkx65\") pod \"barbican-keystone-listener-7cbc6494d6-8n275\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.276571 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82qv7\" (UniqueName: \"kubernetes.io/projected/2a663936-9c21-4231-ac10-2ff34797b3b5-kube-api-access-82qv7\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.276630 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-combined-ca-bundle\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.276689 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data-custom\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.276756 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvfdb\" (UniqueName: \"kubernetes.io/projected/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-kube-api-access-cvfdb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.276804 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-svc\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.276827 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-swift-storage-0\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.276847 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a663936-9c21-4231-ac10-2ff34797b3b5-logs\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.276883 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.277001 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-sb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.277110 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-nb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.277321 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-config\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.334323 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.363344 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.379653 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-config\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.379756 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82qv7\" (UniqueName: \"kubernetes.io/projected/2a663936-9c21-4231-ac10-2ff34797b3b5-kube-api-access-82qv7\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.379789 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-combined-ca-bundle\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.379841 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data-custom\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.379876 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvfdb\" (UniqueName: \"kubernetes.io/projected/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-kube-api-access-cvfdb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.379947 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-svc\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.379969 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-swift-storage-0\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.379994 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a663936-9c21-4231-ac10-2ff34797b3b5-logs\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.380022 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.380058 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-sb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.380093 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-nb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.380612 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-config\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.381057 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-nb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.381304 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-svc\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.381694 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-swift-storage-0\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.381973 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a663936-9c21-4231-ac10-2ff34797b3b5-logs\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.382996 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-sb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.385860 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.387400 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-combined-ca-bundle\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.389715 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data-custom\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.395357 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82qv7\" (UniqueName: \"kubernetes.io/projected/2a663936-9c21-4231-ac10-2ff34797b3b5-kube-api-access-82qv7\") pod \"barbican-api-75db496546-lvnqq\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.405629 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvfdb\" (UniqueName: \"kubernetes.io/projected/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-kube-api-access-cvfdb\") pod \"dnsmasq-dns-548c5895b5-k9hln\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.492134 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.520244 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.796204 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e94e21e9-e1a5-406b-9235-35165cded309","Type":"ContainerStarted","Data":"b4761f2806893419fa049bdae0c7dd84eb9621a5db9d1c6aa5094759cbd94a58"} Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.796494 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e94e21e9-e1a5-406b-9235-35165cded309" containerName="cinder-api-log" containerID="cri-o://e57b27b19e3967d9be9e7a1b859310794e3bc546a933ee59a48b2cc0ef49ea92" gracePeriod=30 Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.796537 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.796927 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e94e21e9-e1a5-406b-9235-35165cded309" containerName="cinder-api" containerID="cri-o://b4761f2806893419fa049bdae0c7dd84eb9621a5db9d1c6aa5094759cbd94a58" gracePeriod=30 Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.802506 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerStarted","Data":"835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69"} Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.807099 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" event={"ID":"20540188-e24e-48b4-acde-2214c8b4d2b9","Type":"ContainerStarted","Data":"52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be"} Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.807239 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" podUID="20540188-e24e-48b4-acde-2214c8b4d2b9" containerName="dnsmasq-dns" containerID="cri-o://52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be" gracePeriod=10 Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.807257 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.881487 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.869879548 podStartE2EDuration="2.869879548s" podCreationTimestamp="2025-10-05 07:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:55.818684176 +0000 UTC m=+1209.701310636" watchObservedRunningTime="2025-10-05 07:12:55.869879548 +0000 UTC m=+1209.752506008" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.914807 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7cbc6494d6-8n275"] Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.924655 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" podStartSLOduration=3.9246342050000003 podStartE2EDuration="3.924634205s" podCreationTimestamp="2025-10-05 07:12:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:55.892425528 +0000 UTC m=+1209.775052008" watchObservedRunningTime="2025-10-05 07:12:55.924634205 +0000 UTC m=+1209.807260655" Oct 05 07:12:55 crc kubenswrapper[4935]: I1005 07:12:55.947497 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57fd77d677-fr4zj"] Oct 05 07:12:55 crc kubenswrapper[4935]: W1005 07:12:55.958464 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5efc4a0_badc_4f8c_a58b_ef0f5e42f93b.slice/crio-38551ca6015da5ff6432e39dd2e14f955aeb6d46439d53443c31bfff7131960d WatchSource:0}: Error finding container 38551ca6015da5ff6432e39dd2e14f955aeb6d46439d53443c31bfff7131960d: Status 404 returned error can't find the container with id 38551ca6015da5ff6432e39dd2e14f955aeb6d46439d53443c31bfff7131960d Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.127373 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-k9hln"] Oct 05 07:12:56 crc kubenswrapper[4935]: W1005 07:12:56.160102 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb4c0762_88dd_404e_8cfc_6ed3860f41b3.slice/crio-26cf71adea4779a7058bfd6cc8b7a2691c2b9ea0046d84682731c6c608f3c4c7 WatchSource:0}: Error finding container 26cf71adea4779a7058bfd6cc8b7a2691c2b9ea0046d84682731c6c608f3c4c7: Status 404 returned error can't find the container with id 26cf71adea4779a7058bfd6cc8b7a2691c2b9ea0046d84682731c6c608f3c4c7 Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.176111 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75db496546-lvnqq"] Oct 05 07:12:56 crc kubenswrapper[4935]: W1005 07:12:56.197084 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a663936_9c21_4231_ac10_2ff34797b3b5.slice/crio-c3882513d73093841f2caecb7d72699f407cfda0cfec118a190e466c1e414f92 WatchSource:0}: Error finding container c3882513d73093841f2caecb7d72699f407cfda0cfec118a190e466c1e414f92: Status 404 returned error can't find the container with id c3882513d73093841f2caecb7d72699f407cfda0cfec118a190e466c1e414f92 Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.264705 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.415415 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-svc\") pod \"20540188-e24e-48b4-acde-2214c8b4d2b9\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.415876 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-config\") pod \"20540188-e24e-48b4-acde-2214c8b4d2b9\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.416204 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-nb\") pod \"20540188-e24e-48b4-acde-2214c8b4d2b9\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.416285 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-sb\") pod \"20540188-e24e-48b4-acde-2214c8b4d2b9\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.416306 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-swift-storage-0\") pod \"20540188-e24e-48b4-acde-2214c8b4d2b9\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.416361 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x68v4\" (UniqueName: \"kubernetes.io/projected/20540188-e24e-48b4-acde-2214c8b4d2b9-kube-api-access-x68v4\") pod \"20540188-e24e-48b4-acde-2214c8b4d2b9\" (UID: \"20540188-e24e-48b4-acde-2214c8b4d2b9\") " Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.426135 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20540188-e24e-48b4-acde-2214c8b4d2b9-kube-api-access-x68v4" (OuterVolumeSpecName: "kube-api-access-x68v4") pod "20540188-e24e-48b4-acde-2214c8b4d2b9" (UID: "20540188-e24e-48b4-acde-2214c8b4d2b9"). InnerVolumeSpecName "kube-api-access-x68v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.477564 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "20540188-e24e-48b4-acde-2214c8b4d2b9" (UID: "20540188-e24e-48b4-acde-2214c8b4d2b9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.479912 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "20540188-e24e-48b4-acde-2214c8b4d2b9" (UID: "20540188-e24e-48b4-acde-2214c8b4d2b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.484783 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-config" (OuterVolumeSpecName: "config") pod "20540188-e24e-48b4-acde-2214c8b4d2b9" (UID: "20540188-e24e-48b4-acde-2214c8b4d2b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.487843 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "20540188-e24e-48b4-acde-2214c8b4d2b9" (UID: "20540188-e24e-48b4-acde-2214c8b4d2b9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.489183 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "20540188-e24e-48b4-acde-2214c8b4d2b9" (UID: "20540188-e24e-48b4-acde-2214c8b4d2b9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.518355 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.518388 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.518397 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.518407 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.518417 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20540188-e24e-48b4-acde-2214c8b4d2b9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.518427 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x68v4\" (UniqueName: \"kubernetes.io/projected/20540188-e24e-48b4-acde-2214c8b4d2b9-kube-api-access-x68v4\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.818202 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75db496546-lvnqq" event={"ID":"2a663936-9c21-4231-ac10-2ff34797b3b5","Type":"ContainerStarted","Data":"ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.818246 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75db496546-lvnqq" event={"ID":"2a663936-9c21-4231-ac10-2ff34797b3b5","Type":"ContainerStarted","Data":"c3882513d73093841f2caecb7d72699f407cfda0cfec118a190e466c1e414f92"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.821046 4935 generic.go:334] "Generic (PLEG): container finished" podID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" containerID="717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab" exitCode=0 Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.821134 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" event={"ID":"eb4c0762-88dd-404e-8cfc-6ed3860f41b3","Type":"ContainerDied","Data":"717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.821159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" event={"ID":"eb4c0762-88dd-404e-8cfc-6ed3860f41b3","Type":"ContainerStarted","Data":"26cf71adea4779a7058bfd6cc8b7a2691c2b9ea0046d84682731c6c608f3c4c7"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.825955 4935 generic.go:334] "Generic (PLEG): container finished" podID="20540188-e24e-48b4-acde-2214c8b4d2b9" containerID="52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be" exitCode=0 Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.826007 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" event={"ID":"20540188-e24e-48b4-acde-2214c8b4d2b9","Type":"ContainerDied","Data":"52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.826032 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" event={"ID":"20540188-e24e-48b4-acde-2214c8b4d2b9","Type":"ContainerDied","Data":"d249b1162e1a38677438f864e5f05a09ca0f14734edc43f4840987d391057dfd"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.826049 4935 scope.go:117] "RemoveContainer" containerID="52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.826160 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b76d77dc-clsjp" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.836716 4935 generic.go:334] "Generic (PLEG): container finished" podID="e94e21e9-e1a5-406b-9235-35165cded309" containerID="e57b27b19e3967d9be9e7a1b859310794e3bc546a933ee59a48b2cc0ef49ea92" exitCode=143 Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.836779 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e94e21e9-e1a5-406b-9235-35165cded309","Type":"ContainerDied","Data":"e57b27b19e3967d9be9e7a1b859310794e3bc546a933ee59a48b2cc0ef49ea92"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.840880 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f66f223c-e87c-4314-957f-bd7c55dafc8f","Type":"ContainerStarted","Data":"a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.850172 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerStarted","Data":"39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.856406 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57fd77d677-fr4zj" event={"ID":"c9b2e8af-97b0-4103-a241-ca473f299043","Type":"ContainerStarted","Data":"85dd90408525c28687ced0ae2ac7d321bf695ebcedcee982cb394973277cd908"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.858981 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" event={"ID":"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b","Type":"ContainerStarted","Data":"38551ca6015da5ff6432e39dd2e14f955aeb6d46439d53443c31bfff7131960d"} Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.930857 4935 scope.go:117] "RemoveContainer" containerID="ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3" Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.981124 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77b76d77dc-clsjp"] Oct 05 07:12:56 crc kubenswrapper[4935]: I1005 07:12:56.999593 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77b76d77dc-clsjp"] Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.021238 4935 scope.go:117] "RemoveContainer" containerID="52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be" Oct 05 07:12:57 crc kubenswrapper[4935]: E1005 07:12:57.021633 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be\": container with ID starting with 52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be not found: ID does not exist" containerID="52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.021685 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be"} err="failed to get container status \"52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be\": rpc error: code = NotFound desc = could not find container \"52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be\": container with ID starting with 52661ad06a99d49827bc2c6e6b953fe9a4f7028d3b9c3afe9ffe0898df4fd5be not found: ID does not exist" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.021709 4935 scope.go:117] "RemoveContainer" containerID="ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3" Oct 05 07:12:57 crc kubenswrapper[4935]: E1005 07:12:57.022122 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3\": container with ID starting with ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3 not found: ID does not exist" containerID="ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.022155 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3"} err="failed to get container status \"ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3\": rpc error: code = NotFound desc = could not find container \"ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3\": container with ID starting with ed9cf747435e560cc836666b0fb9c2b3650a5f2015615bcdbe94ab7de3c040d3 not found: ID does not exist" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.873548 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerStarted","Data":"b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c"} Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.884189 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f66f223c-e87c-4314-957f-bd7c55dafc8f","Type":"ContainerStarted","Data":"b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c"} Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.894738 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75db496546-lvnqq" event={"ID":"2a663936-9c21-4231-ac10-2ff34797b3b5","Type":"ContainerStarted","Data":"fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23"} Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.895149 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.895245 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.898147 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" event={"ID":"eb4c0762-88dd-404e-8cfc-6ed3860f41b3","Type":"ContainerStarted","Data":"db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92"} Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.898292 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.936412 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.841231272 podStartE2EDuration="5.936393841s" podCreationTimestamp="2025-10-05 07:12:52 +0000 UTC" firstStartedPulling="2025-10-05 07:12:53.760962276 +0000 UTC m=+1207.643588736" lastFinishedPulling="2025-10-05 07:12:54.856124845 +0000 UTC m=+1208.738751305" observedRunningTime="2025-10-05 07:12:57.912908267 +0000 UTC m=+1211.795534727" watchObservedRunningTime="2025-10-05 07:12:57.936393841 +0000 UTC m=+1211.819020301" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.941153 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-75db496546-lvnqq" podStartSLOduration=2.941128467 podStartE2EDuration="2.941128467s" podCreationTimestamp="2025-10-05 07:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:57.929616961 +0000 UTC m=+1211.812243421" watchObservedRunningTime="2025-10-05 07:12:57.941128467 +0000 UTC m=+1211.823754947" Oct 05 07:12:57 crc kubenswrapper[4935]: I1005 07:12:57.999549 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" podStartSLOduration=2.999527601 podStartE2EDuration="2.999527601s" podCreationTimestamp="2025-10-05 07:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:12:57.959441605 +0000 UTC m=+1211.842068105" watchObservedRunningTime="2025-10-05 07:12:57.999527601 +0000 UTC m=+1211.882154061" Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.271380 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.788389 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20540188-e24e-48b4-acde-2214c8b4d2b9" path="/var/lib/kubelet/pods/20540188-e24e-48b4-acde-2214c8b4d2b9/volumes" Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.915380 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" event={"ID":"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b","Type":"ContainerStarted","Data":"9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00"} Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.915422 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" event={"ID":"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b","Type":"ContainerStarted","Data":"8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a"} Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.919551 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerStarted","Data":"fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e"} Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.919780 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.928250 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57fd77d677-fr4zj" event={"ID":"c9b2e8af-97b0-4103-a241-ca473f299043","Type":"ContainerStarted","Data":"f1adbd1314cf52e09287126eae2dabba2ce549b2976316b1a565b87702fb1644"} Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.928394 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57fd77d677-fr4zj" event={"ID":"c9b2e8af-97b0-4103-a241-ca473f299043","Type":"ContainerStarted","Data":"a5103fe4899d57404d97b7ee8251771675c6a622925145cac9caa880ee1b5c91"} Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.948908 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" podStartSLOduration=2.971820708 podStartE2EDuration="4.948870071s" podCreationTimestamp="2025-10-05 07:12:54 +0000 UTC" firstStartedPulling="2025-10-05 07:12:55.967500575 +0000 UTC m=+1209.850127035" lastFinishedPulling="2025-10-05 07:12:57.944549928 +0000 UTC m=+1211.827176398" observedRunningTime="2025-10-05 07:12:58.94735132 +0000 UTC m=+1212.829977780" watchObservedRunningTime="2025-10-05 07:12:58.948870071 +0000 UTC m=+1212.831496531" Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.975871 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-57fd77d677-fr4zj" podStartSLOduration=3.008916064 podStartE2EDuration="4.975854768s" podCreationTimestamp="2025-10-05 07:12:54 +0000 UTC" firstStartedPulling="2025-10-05 07:12:55.98008757 +0000 UTC m=+1209.862714030" lastFinishedPulling="2025-10-05 07:12:57.947026254 +0000 UTC m=+1211.829652734" observedRunningTime="2025-10-05 07:12:58.970049534 +0000 UTC m=+1212.852675994" watchObservedRunningTime="2025-10-05 07:12:58.975854768 +0000 UTC m=+1212.858481228" Oct 05 07:12:58 crc kubenswrapper[4935]: I1005 07:12:58.996232 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.03331474 podStartE2EDuration="6.99621236s" podCreationTimestamp="2025-10-05 07:12:52 +0000 UTC" firstStartedPulling="2025-10-05 07:12:53.994342946 +0000 UTC m=+1207.876969406" lastFinishedPulling="2025-10-05 07:12:57.957240566 +0000 UTC m=+1211.839867026" observedRunningTime="2025-10-05 07:12:58.987607101 +0000 UTC m=+1212.870233561" watchObservedRunningTime="2025-10-05 07:12:58.99621236 +0000 UTC m=+1212.878838820" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.164934 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-74989f54d8-bw287"] Oct 05 07:12:59 crc kubenswrapper[4935]: E1005 07:12:59.165387 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20540188-e24e-48b4-acde-2214c8b4d2b9" containerName="dnsmasq-dns" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.165411 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="20540188-e24e-48b4-acde-2214c8b4d2b9" containerName="dnsmasq-dns" Oct 05 07:12:59 crc kubenswrapper[4935]: E1005 07:12:59.165445 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20540188-e24e-48b4-acde-2214c8b4d2b9" containerName="init" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.165454 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="20540188-e24e-48b4-acde-2214c8b4d2b9" containerName="init" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.165647 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="20540188-e24e-48b4-acde-2214c8b4d2b9" containerName="dnsmasq-dns" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.166597 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.168761 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.169001 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.182735 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74989f54d8-bw287"] Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.271833 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.271938 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-public-tls-certs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.272382 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-combined-ca-bundle\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.272505 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-internal-tls-certs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.272688 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8dcg\" (UniqueName: \"kubernetes.io/projected/3c422d4a-a835-443a-9c6d-3161bc80c1aa-kube-api-access-d8dcg\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.272773 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c422d4a-a835-443a-9c6d-3161bc80c1aa-logs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.273034 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data-custom\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.374778 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-combined-ca-bundle\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.374857 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-internal-tls-certs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.375008 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8dcg\" (UniqueName: \"kubernetes.io/projected/3c422d4a-a835-443a-9c6d-3161bc80c1aa-kube-api-access-d8dcg\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.375045 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c422d4a-a835-443a-9c6d-3161bc80c1aa-logs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.375115 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data-custom\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.375165 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.375189 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-public-tls-certs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.375561 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c422d4a-a835-443a-9c6d-3161bc80c1aa-logs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.381440 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data-custom\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.381486 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-internal-tls-certs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.381873 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-public-tls-certs\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.385799 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.391962 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-combined-ca-bundle\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.398714 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8dcg\" (UniqueName: \"kubernetes.io/projected/3c422d4a-a835-443a-9c6d-3161bc80c1aa-kube-api-access-d8dcg\") pod \"barbican-api-74989f54d8-bw287\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.484656 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:12:59 crc kubenswrapper[4935]: I1005 07:12:59.993665 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74989f54d8-bw287"] Oct 05 07:13:00 crc kubenswrapper[4935]: I1005 07:13:00.946426 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74989f54d8-bw287" event={"ID":"3c422d4a-a835-443a-9c6d-3161bc80c1aa","Type":"ContainerStarted","Data":"119888a24ef912962cb413c88d9903d21aad373b3de77a30cfb8ecc482ff8f75"} Oct 05 07:13:00 crc kubenswrapper[4935]: I1005 07:13:00.946823 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74989f54d8-bw287" event={"ID":"3c422d4a-a835-443a-9c6d-3161bc80c1aa","Type":"ContainerStarted","Data":"9b2b88228c11fcf40bd73c44aadf5f4a93619dfc6b49a5c328edcf98f4d93a7e"} Oct 05 07:13:00 crc kubenswrapper[4935]: I1005 07:13:00.946834 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74989f54d8-bw287" event={"ID":"3c422d4a-a835-443a-9c6d-3161bc80c1aa","Type":"ContainerStarted","Data":"c584480b35bceb761270e71b6c5a36b7d7a92b0e54a1575e6dbcf0164b50ed3b"} Oct 05 07:13:00 crc kubenswrapper[4935]: I1005 07:13:00.947045 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:13:00 crc kubenswrapper[4935]: I1005 07:13:00.947113 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:13:00 crc kubenswrapper[4935]: I1005 07:13:00.971952 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-74989f54d8-bw287" podStartSLOduration=1.9719298570000001 podStartE2EDuration="1.971929857s" podCreationTimestamp="2025-10-05 07:12:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:13:00.963107312 +0000 UTC m=+1214.845733772" watchObservedRunningTime="2025-10-05 07:13:00.971929857 +0000 UTC m=+1214.854556317" Oct 05 07:13:03 crc kubenswrapper[4935]: I1005 07:13:03.496570 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 07:13:03 crc kubenswrapper[4935]: I1005 07:13:03.567648 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:13:03 crc kubenswrapper[4935]: I1005 07:13:03.973972 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerName="cinder-scheduler" containerID="cri-o://a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8" gracePeriod=30 Oct 05 07:13:03 crc kubenswrapper[4935]: I1005 07:13:03.974156 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerName="probe" containerID="cri-o://b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c" gracePeriod=30 Oct 05 07:13:04 crc kubenswrapper[4935]: I1005 07:13:04.986983 4935 generic.go:334] "Generic (PLEG): container finished" podID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerID="b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c" exitCode=0 Oct 05 07:13:04 crc kubenswrapper[4935]: I1005 07:13:04.987224 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f66f223c-e87c-4314-957f-bd7c55dafc8f","Type":"ContainerDied","Data":"b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c"} Oct 05 07:13:05 crc kubenswrapper[4935]: I1005 07:13:05.480086 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 07:13:05 crc kubenswrapper[4935]: I1005 07:13:05.494796 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:13:05 crc kubenswrapper[4935]: I1005 07:13:05.626313 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc97d9d55-72qqm"] Oct 05 07:13:05 crc kubenswrapper[4935]: I1005 07:13:05.626595 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" podUID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerName="dnsmasq-dns" containerID="cri-o://2f9de2f2ec63a3b30bf4e29252e9801a9c005e253ce47dd65f96d3a34902f706" gracePeriod=10 Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.007264 4935 generic.go:334] "Generic (PLEG): container finished" podID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerID="2f9de2f2ec63a3b30bf4e29252e9801a9c005e253ce47dd65f96d3a34902f706" exitCode=0 Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.007633 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" event={"ID":"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79","Type":"ContainerDied","Data":"2f9de2f2ec63a3b30bf4e29252e9801a9c005e253ce47dd65f96d3a34902f706"} Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.307261 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.320058 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.431539 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-swift-storage-0\") pod \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.431602 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cscw6\" (UniqueName: \"kubernetes.io/projected/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-kube-api-access-cscw6\") pod \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.431653 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-sb\") pod \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.431735 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-svc\") pod \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.431761 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-nb\") pod \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.431818 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-config\") pod \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\" (UID: \"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79\") " Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.470074 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-kube-api-access-cscw6" (OuterVolumeSpecName: "kube-api-access-cscw6") pod "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" (UID: "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79"). InnerVolumeSpecName "kube-api-access-cscw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.534817 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cscw6\" (UniqueName: \"kubernetes.io/projected/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-kube-api-access-cscw6\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.547605 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" (UID: "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.572066 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" (UID: "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.574455 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-config" (OuterVolumeSpecName: "config") pod "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" (UID: "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.575388 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" (UID: "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.580351 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" (UID: "f2e01050-b5d1-496c-b24e-a2d1cdb4cb79"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.636344 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.636380 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.636391 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.636404 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:06 crc kubenswrapper[4935]: I1005 07:13:06.636413 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.017997 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" event={"ID":"f2e01050-b5d1-496c-b24e-a2d1cdb4cb79","Type":"ContainerDied","Data":"e03a97531abbe05ee0a2d911cef019568273b320014dfebbc03442aca7b4dac2"} Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.018042 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.018052 4935 scope.go:117] "RemoveContainer" containerID="2f9de2f2ec63a3b30bf4e29252e9801a9c005e253ce47dd65f96d3a34902f706" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.047337 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc97d9d55-72qqm"] Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.055886 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc97d9d55-72qqm"] Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.059037 4935 scope.go:117] "RemoveContainer" containerID="51b2ef43d64a48354781cfd1e67ed1fa7c882a3527b2e0cffe22d77f0a02679f" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.424638 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.432037 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.828047 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.871871 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.961185 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-scripts\") pod \"f66f223c-e87c-4314-957f-bd7c55dafc8f\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.961258 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f66f223c-e87c-4314-957f-bd7c55dafc8f-etc-machine-id\") pod \"f66f223c-e87c-4314-957f-bd7c55dafc8f\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.961348 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8cpx\" (UniqueName: \"kubernetes.io/projected/f66f223c-e87c-4314-957f-bd7c55dafc8f-kube-api-access-v8cpx\") pod \"f66f223c-e87c-4314-957f-bd7c55dafc8f\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.961376 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data\") pod \"f66f223c-e87c-4314-957f-bd7c55dafc8f\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.961408 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data-custom\") pod \"f66f223c-e87c-4314-957f-bd7c55dafc8f\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.961446 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-combined-ca-bundle\") pod \"f66f223c-e87c-4314-957f-bd7c55dafc8f\" (UID: \"f66f223c-e87c-4314-957f-bd7c55dafc8f\") " Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.962258 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f66f223c-e87c-4314-957f-bd7c55dafc8f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f66f223c-e87c-4314-957f-bd7c55dafc8f" (UID: "f66f223c-e87c-4314-957f-bd7c55dafc8f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.967109 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f66f223c-e87c-4314-957f-bd7c55dafc8f" (UID: "f66f223c-e87c-4314-957f-bd7c55dafc8f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.967172 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f66f223c-e87c-4314-957f-bd7c55dafc8f-kube-api-access-v8cpx" (OuterVolumeSpecName: "kube-api-access-v8cpx") pod "f66f223c-e87c-4314-957f-bd7c55dafc8f" (UID: "f66f223c-e87c-4314-957f-bd7c55dafc8f"). InnerVolumeSpecName "kube-api-access-v8cpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:07 crc kubenswrapper[4935]: I1005 07:13:07.993043 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-scripts" (OuterVolumeSpecName: "scripts") pod "f66f223c-e87c-4314-957f-bd7c55dafc8f" (UID: "f66f223c-e87c-4314-957f-bd7c55dafc8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.023324 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f66f223c-e87c-4314-957f-bd7c55dafc8f" (UID: "f66f223c-e87c-4314-957f-bd7c55dafc8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.043624 4935 generic.go:334] "Generic (PLEG): container finished" podID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerID="a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8" exitCode=0 Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.043691 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f66f223c-e87c-4314-957f-bd7c55dafc8f","Type":"ContainerDied","Data":"a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8"} Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.043701 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.043761 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f66f223c-e87c-4314-957f-bd7c55dafc8f","Type":"ContainerDied","Data":"546d24e9310c8d9cd95ff13ba5c5d2410e8fd137dfabbe554bb38aa898d2367d"} Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.043785 4935 scope.go:117] "RemoveContainer" containerID="b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.066075 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8cpx\" (UniqueName: \"kubernetes.io/projected/f66f223c-e87c-4314-957f-bd7c55dafc8f-kube-api-access-v8cpx\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.066113 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.066124 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.066136 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.066147 4935 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f66f223c-e87c-4314-957f-bd7c55dafc8f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.085998 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data" (OuterVolumeSpecName: "config-data") pod "f66f223c-e87c-4314-957f-bd7c55dafc8f" (UID: "f66f223c-e87c-4314-957f-bd7c55dafc8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.098636 4935 scope.go:117] "RemoveContainer" containerID="a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.117966 4935 scope.go:117] "RemoveContainer" containerID="b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c" Oct 05 07:13:08 crc kubenswrapper[4935]: E1005 07:13:08.118616 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c\": container with ID starting with b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c not found: ID does not exist" containerID="b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.118703 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c"} err="failed to get container status \"b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c\": rpc error: code = NotFound desc = could not find container \"b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c\": container with ID starting with b2c3dc39d2866e67be83011f2971190875a3b175f3b6cc14e080cf742aef3f2c not found: ID does not exist" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.118762 4935 scope.go:117] "RemoveContainer" containerID="a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8" Oct 05 07:13:08 crc kubenswrapper[4935]: E1005 07:13:08.119246 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8\": container with ID starting with a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8 not found: ID does not exist" containerID="a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.119286 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8"} err="failed to get container status \"a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8\": rpc error: code = NotFound desc = could not find container \"a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8\": container with ID starting with a1cf1ef7233d759dd4908a815db6e537db6bf98bff0e870480503a8aea9600d8 not found: ID does not exist" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.168145 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f66f223c-e87c-4314-957f-bd7c55dafc8f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.183077 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.394581 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.415125 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.415880 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:13:08 crc kubenswrapper[4935]: E1005 07:13:08.416785 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerName="probe" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.416821 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerName="probe" Oct 05 07:13:08 crc kubenswrapper[4935]: E1005 07:13:08.416843 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerName="cinder-scheduler" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.416856 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerName="cinder-scheduler" Oct 05 07:13:08 crc kubenswrapper[4935]: E1005 07:13:08.416918 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerName="dnsmasq-dns" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.416933 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerName="dnsmasq-dns" Oct 05 07:13:08 crc kubenswrapper[4935]: E1005 07:13:08.416968 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerName="init" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.416983 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerName="init" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.417328 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerName="probe" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.417393 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" containerName="cinder-scheduler" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.417423 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerName="dnsmasq-dns" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.419501 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.424455 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.434947 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.476006 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9acce3fe-5a42-49e7-b3e7-a72f01af5736-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.476139 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-scripts\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.476914 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.476990 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rjgb\" (UniqueName: \"kubernetes.io/projected/9acce3fe-5a42-49e7-b3e7-a72f01af5736-kube-api-access-2rjgb\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.477083 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.477302 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.548351 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.580478 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.580567 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rjgb\" (UniqueName: \"kubernetes.io/projected/9acce3fe-5a42-49e7-b3e7-a72f01af5736-kube-api-access-2rjgb\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.580662 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.580820 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.580885 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9acce3fe-5a42-49e7-b3e7-a72f01af5736-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.581055 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-scripts\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.583017 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9acce3fe-5a42-49e7-b3e7-a72f01af5736-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.585807 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-scripts\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.585972 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.586309 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.586979 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.621572 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rjgb\" (UniqueName: \"kubernetes.io/projected/9acce3fe-5a42-49e7-b3e7-a72f01af5736-kube-api-access-2rjgb\") pod \"cinder-scheduler-0\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.641137 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.658117 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56b79f594d-q76hh"] Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.658466 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56b79f594d-q76hh" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerName="neutron-api" containerID="cri-o://ad39ce747967bfbf8195ccbec83da9a5e17c907f4a3c9ca8bd0259457636a718" gracePeriod=30 Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.658566 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56b79f594d-q76hh" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerName="neutron-httpd" containerID="cri-o://76684d743646083119a5adf79f28378bb03608f41db9fee37a119f5a9eba8108" gracePeriod=30 Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.775542 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.801314 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" path="/var/lib/kubelet/pods/f2e01050-b5d1-496c-b24e-a2d1cdb4cb79/volumes" Oct 05 07:13:08 crc kubenswrapper[4935]: I1005 07:13:08.802463 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f66f223c-e87c-4314-957f-bd7c55dafc8f" path="/var/lib/kubelet/pods/f66f223c-e87c-4314-957f-bd7c55dafc8f/volumes" Oct 05 07:13:09 crc kubenswrapper[4935]: I1005 07:13:09.069335 4935 generic.go:334] "Generic (PLEG): container finished" podID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerID="76684d743646083119a5adf79f28378bb03608f41db9fee37a119f5a9eba8108" exitCode=0 Oct 05 07:13:09 crc kubenswrapper[4935]: I1005 07:13:09.069423 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56b79f594d-q76hh" event={"ID":"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8","Type":"ContainerDied","Data":"76684d743646083119a5adf79f28378bb03608f41db9fee37a119f5a9eba8108"} Oct 05 07:13:09 crc kubenswrapper[4935]: I1005 07:13:09.442134 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:13:09 crc kubenswrapper[4935]: W1005 07:13:09.453715 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9acce3fe_5a42_49e7_b3e7_a72f01af5736.slice/crio-d5a662fd065201895283c03ffbb32808389eddd8ab036e44f1677d1a8f7b9801 WatchSource:0}: Error finding container d5a662fd065201895283c03ffbb32808389eddd8ab036e44f1677d1a8f7b9801: Status 404 returned error can't find the container with id d5a662fd065201895283c03ffbb32808389eddd8ab036e44f1677d1a8f7b9801 Oct 05 07:13:10 crc kubenswrapper[4935]: I1005 07:13:10.085644 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9acce3fe-5a42-49e7-b3e7-a72f01af5736","Type":"ContainerStarted","Data":"d5a662fd065201895283c03ffbb32808389eddd8ab036e44f1677d1a8f7b9801"} Oct 05 07:13:10 crc kubenswrapper[4935]: I1005 07:13:10.985119 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bc97d9d55-72qqm" podUID="f2e01050-b5d1-496c-b24e-a2d1cdb4cb79" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: i/o timeout" Oct 05 07:13:11 crc kubenswrapper[4935]: I1005 07:13:11.028643 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:13:11 crc kubenswrapper[4935]: I1005 07:13:11.051286 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:13:11 crc kubenswrapper[4935]: I1005 07:13:11.097611 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9acce3fe-5a42-49e7-b3e7-a72f01af5736","Type":"ContainerStarted","Data":"c21841257bca2cda12233fe43bab1edd250a07081b7783d9448defce60d379a8"} Oct 05 07:13:11 crc kubenswrapper[4935]: I1005 07:13:11.097645 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9acce3fe-5a42-49e7-b3e7-a72f01af5736","Type":"ContainerStarted","Data":"24b5a0ee97d3993b21bd1da36e2ba46a94d38c77257bf5f70b85ea35748a294c"} Oct 05 07:13:11 crc kubenswrapper[4935]: I1005 07:13:11.110330 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-75db496546-lvnqq"] Oct 05 07:13:11 crc kubenswrapper[4935]: I1005 07:13:11.110530 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-75db496546-lvnqq" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerName="barbican-api-log" containerID="cri-o://ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0" gracePeriod=30 Oct 05 07:13:11 crc kubenswrapper[4935]: I1005 07:13:11.110626 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-75db496546-lvnqq" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerName="barbican-api" containerID="cri-o://fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23" gracePeriod=30 Oct 05 07:13:11 crc kubenswrapper[4935]: I1005 07:13:11.127569 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.127544983 podStartE2EDuration="3.127544983s" podCreationTimestamp="2025-10-05 07:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:13:11.121737209 +0000 UTC m=+1225.004363659" watchObservedRunningTime="2025-10-05 07:13:11.127544983 +0000 UTC m=+1225.010171443" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.120089 4935 generic.go:334] "Generic (PLEG): container finished" podID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerID="ad39ce747967bfbf8195ccbec83da9a5e17c907f4a3c9ca8bd0259457636a718" exitCode=0 Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.120284 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56b79f594d-q76hh" event={"ID":"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8","Type":"ContainerDied","Data":"ad39ce747967bfbf8195ccbec83da9a5e17c907f4a3c9ca8bd0259457636a718"} Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.122118 4935 generic.go:334] "Generic (PLEG): container finished" podID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerID="ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0" exitCode=143 Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.122354 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75db496546-lvnqq" event={"ID":"2a663936-9c21-4231-ac10-2ff34797b3b5","Type":"ContainerDied","Data":"ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0"} Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.227223 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.228365 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.236743 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-b6wqw" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.236941 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.237057 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.249038 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.283284 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config-secret\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.283369 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.283433 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.283456 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rtjh\" (UniqueName: \"kubernetes.io/projected/abe76393-92f2-4164-a549-5f75ced24fb3-kube-api-access-8rtjh\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.366439 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.384439 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config-secret\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.384520 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.384578 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.384598 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rtjh\" (UniqueName: \"kubernetes.io/projected/abe76393-92f2-4164-a549-5f75ced24fb3-kube-api-access-8rtjh\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.385726 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.396793 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config-secret\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.404495 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.406917 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rtjh\" (UniqueName: \"kubernetes.io/projected/abe76393-92f2-4164-a549-5f75ced24fb3-kube-api-access-8rtjh\") pod \"openstackclient\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.485458 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-ovndb-tls-certs\") pod \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.485772 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-combined-ca-bundle\") pod \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.486082 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-httpd-config\") pod \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.486227 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrthr\" (UniqueName: \"kubernetes.io/projected/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-kube-api-access-xrthr\") pod \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.486346 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-config\") pod \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\" (UID: \"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8\") " Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.488973 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-kube-api-access-xrthr" (OuterVolumeSpecName: "kube-api-access-xrthr") pod "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" (UID: "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8"). InnerVolumeSpecName "kube-api-access-xrthr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.513135 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" (UID: "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.548257 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-config" (OuterVolumeSpecName: "config") pod "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" (UID: "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.568974 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" (UID: "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.568995 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.587936 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.587968 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.587978 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrthr\" (UniqueName: \"kubernetes.io/projected/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-kube-api-access-xrthr\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.587987 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.602560 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" (UID: "eeb801c8-5f9c-4bfa-8f21-80a046fd04a8"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:12 crc kubenswrapper[4935]: I1005 07:13:12.689177 4935 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.017878 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.131766 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56b79f594d-q76hh" event={"ID":"eeb801c8-5f9c-4bfa-8f21-80a046fd04a8","Type":"ContainerDied","Data":"0bd8ecbea4c2c8c483b0c8640819f1183b03355701c0b85985b269fe36e28c68"} Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.131854 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56b79f594d-q76hh" Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.133296 4935 scope.go:117] "RemoveContainer" containerID="76684d743646083119a5adf79f28378bb03608f41db9fee37a119f5a9eba8108" Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.133189 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"abe76393-92f2-4164-a549-5f75ced24fb3","Type":"ContainerStarted","Data":"1bce92b740c348f19dcc8c8378e02e645109f55c01071b2a268b74ea074790e7"} Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.159040 4935 scope.go:117] "RemoveContainer" containerID="ad39ce747967bfbf8195ccbec83da9a5e17c907f4a3c9ca8bd0259457636a718" Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.163203 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56b79f594d-q76hh"] Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.174857 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-56b79f594d-q76hh"] Oct 05 07:13:13 crc kubenswrapper[4935]: I1005 07:13:13.776764 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.717236 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.789364 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" path="/var/lib/kubelet/pods/eeb801c8-5f9c-4bfa-8f21-80a046fd04a8/volumes" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.837941 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data-custom\") pod \"2a663936-9c21-4231-ac10-2ff34797b3b5\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.838391 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data\") pod \"2a663936-9c21-4231-ac10-2ff34797b3b5\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.838432 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82qv7\" (UniqueName: \"kubernetes.io/projected/2a663936-9c21-4231-ac10-2ff34797b3b5-kube-api-access-82qv7\") pod \"2a663936-9c21-4231-ac10-2ff34797b3b5\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.838485 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a663936-9c21-4231-ac10-2ff34797b3b5-logs\") pod \"2a663936-9c21-4231-ac10-2ff34797b3b5\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.838609 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-combined-ca-bundle\") pod \"2a663936-9c21-4231-ac10-2ff34797b3b5\" (UID: \"2a663936-9c21-4231-ac10-2ff34797b3b5\") " Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.839013 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a663936-9c21-4231-ac10-2ff34797b3b5-logs" (OuterVolumeSpecName: "logs") pod "2a663936-9c21-4231-ac10-2ff34797b3b5" (UID: "2a663936-9c21-4231-ac10-2ff34797b3b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.839286 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a663936-9c21-4231-ac10-2ff34797b3b5-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.843842 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2a663936-9c21-4231-ac10-2ff34797b3b5" (UID: "2a663936-9c21-4231-ac10-2ff34797b3b5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.858182 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a663936-9c21-4231-ac10-2ff34797b3b5-kube-api-access-82qv7" (OuterVolumeSpecName: "kube-api-access-82qv7") pod "2a663936-9c21-4231-ac10-2ff34797b3b5" (UID: "2a663936-9c21-4231-ac10-2ff34797b3b5"). InnerVolumeSpecName "kube-api-access-82qv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.898523 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data" (OuterVolumeSpecName: "config-data") pod "2a663936-9c21-4231-ac10-2ff34797b3b5" (UID: "2a663936-9c21-4231-ac10-2ff34797b3b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.923132 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a663936-9c21-4231-ac10-2ff34797b3b5" (UID: "2a663936-9c21-4231-ac10-2ff34797b3b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.941239 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.941301 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82qv7\" (UniqueName: \"kubernetes.io/projected/2a663936-9c21-4231-ac10-2ff34797b3b5-kube-api-access-82qv7\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.941315 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:14 crc kubenswrapper[4935]: I1005 07:13:14.941329 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a663936-9c21-4231-ac10-2ff34797b3b5-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.158325 4935 generic.go:334] "Generic (PLEG): container finished" podID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerID="fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23" exitCode=0 Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.158485 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75db496546-lvnqq" event={"ID":"2a663936-9c21-4231-ac10-2ff34797b3b5","Type":"ContainerDied","Data":"fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23"} Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.159184 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75db496546-lvnqq" event={"ID":"2a663936-9c21-4231-ac10-2ff34797b3b5","Type":"ContainerDied","Data":"c3882513d73093841f2caecb7d72699f407cfda0cfec118a190e466c1e414f92"} Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.159260 4935 scope.go:117] "RemoveContainer" containerID="fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.158579 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75db496546-lvnqq" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.190907 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-75db496546-lvnqq"] Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.197079 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-75db496546-lvnqq"] Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.198609 4935 scope.go:117] "RemoveContainer" containerID="ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.226052 4935 scope.go:117] "RemoveContainer" containerID="fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23" Oct 05 07:13:15 crc kubenswrapper[4935]: E1005 07:13:15.226653 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23\": container with ID starting with fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23 not found: ID does not exist" containerID="fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.226780 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23"} err="failed to get container status \"fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23\": rpc error: code = NotFound desc = could not find container \"fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23\": container with ID starting with fd3c1a01bcfa392a0049ae7eccc66db9f624bfd99226f67607e1bafbc992ef23 not found: ID does not exist" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.226867 4935 scope.go:117] "RemoveContainer" containerID="ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0" Oct 05 07:13:15 crc kubenswrapper[4935]: E1005 07:13:15.227223 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0\": container with ID starting with ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0 not found: ID does not exist" containerID="ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.227308 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0"} err="failed to get container status \"ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0\": rpc error: code = NotFound desc = could not find container \"ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0\": container with ID starting with ceeaba0c4a3fa725dc95dde1fdc981a7901cbf23449ae27be6c3467f8e534ac0 not found: ID does not exist" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.922760 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6f6d895595-jv76b"] Oct 05 07:13:15 crc kubenswrapper[4935]: E1005 07:13:15.923287 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerName="barbican-api" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.923304 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerName="barbican-api" Oct 05 07:13:15 crc kubenswrapper[4935]: E1005 07:13:15.923327 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerName="neutron-httpd" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.923334 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerName="neutron-httpd" Oct 05 07:13:15 crc kubenswrapper[4935]: E1005 07:13:15.923355 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerName="barbican-api-log" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.923364 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerName="barbican-api-log" Oct 05 07:13:15 crc kubenswrapper[4935]: E1005 07:13:15.923380 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerName="neutron-api" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.923387 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerName="neutron-api" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.924192 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerName="neutron-httpd" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.924234 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerName="barbican-api" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.924243 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" containerName="barbican-api-log" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.924252 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeb801c8-5f9c-4bfa-8f21-80a046fd04a8" containerName="neutron-api" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.925248 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.932421 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.932913 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.933023 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 05 07:13:15 crc kubenswrapper[4935]: I1005 07:13:15.935841 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6f6d895595-jv76b"] Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.058831 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-config-data\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.058908 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rngf\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-kube-api-access-8rngf\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.058932 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-run-httpd\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.059141 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-etc-swift\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.059177 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-internal-tls-certs\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.059392 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-log-httpd\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.059451 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-public-tls-certs\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.059481 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-combined-ca-bundle\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.161386 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-etc-swift\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.161422 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-internal-tls-certs\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.161469 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-log-httpd\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.161509 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-public-tls-certs\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.161527 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-combined-ca-bundle\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.161575 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-config-data\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.161604 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rngf\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-kube-api-access-8rngf\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.161624 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-run-httpd\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.162198 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-log-httpd\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.162228 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-run-httpd\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.167938 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-public-tls-certs\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.168573 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-config-data\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.169397 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-etc-swift\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.169636 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-combined-ca-bundle\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.183526 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-internal-tls-certs\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.189519 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rngf\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-kube-api-access-8rngf\") pod \"swift-proxy-6f6d895595-jv76b\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.245532 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.791753 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a663936-9c21-4231-ac10-2ff34797b3b5" path="/var/lib/kubelet/pods/2a663936-9c21-4231-ac10-2ff34797b3b5/volumes" Oct 05 07:13:16 crc kubenswrapper[4935]: I1005 07:13:16.847992 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6f6d895595-jv76b"] Oct 05 07:13:17 crc kubenswrapper[4935]: I1005 07:13:17.530117 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:17 crc kubenswrapper[4935]: I1005 07:13:17.531065 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="ceilometer-central-agent" containerID="cri-o://835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69" gracePeriod=30 Oct 05 07:13:17 crc kubenswrapper[4935]: I1005 07:13:17.531091 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="proxy-httpd" containerID="cri-o://fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e" gracePeriod=30 Oct 05 07:13:17 crc kubenswrapper[4935]: I1005 07:13:17.531179 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="sg-core" containerID="cri-o://b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c" gracePeriod=30 Oct 05 07:13:17 crc kubenswrapper[4935]: I1005 07:13:17.531217 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="ceilometer-notification-agent" containerID="cri-o://39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c" gracePeriod=30 Oct 05 07:13:17 crc kubenswrapper[4935]: I1005 07:13:17.538064 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.153:3000/\": EOF" Oct 05 07:13:18 crc kubenswrapper[4935]: I1005 07:13:18.208356 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerID="fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e" exitCode=0 Oct 05 07:13:18 crc kubenswrapper[4935]: I1005 07:13:18.208382 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerID="b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c" exitCode=2 Oct 05 07:13:18 crc kubenswrapper[4935]: I1005 07:13:18.208391 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerID="835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69" exitCode=0 Oct 05 07:13:18 crc kubenswrapper[4935]: I1005 07:13:18.208410 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerDied","Data":"fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e"} Oct 05 07:13:18 crc kubenswrapper[4935]: I1005 07:13:18.208434 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerDied","Data":"b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c"} Oct 05 07:13:18 crc kubenswrapper[4935]: I1005 07:13:18.208444 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerDied","Data":"835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69"} Oct 05 07:13:19 crc kubenswrapper[4935]: I1005 07:13:19.004489 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 07:13:21 crc kubenswrapper[4935]: W1005 07:13:21.725458 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51e99221_24f0_4cbe_bb3c_4e1e70acd4ba.slice/crio-65e13bd44278107300790a93bc76ecae382f285e1385cee5ef8b18b0a1bbed0b WatchSource:0}: Error finding container 65e13bd44278107300790a93bc76ecae382f285e1385cee5ef8b18b0a1bbed0b: Status 404 returned error can't find the container with id 65e13bd44278107300790a93bc76ecae382f285e1385cee5ef8b18b0a1bbed0b Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.245086 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"abe76393-92f2-4164-a549-5f75ced24fb3","Type":"ContainerStarted","Data":"1d9ad091dc634c8651d67b5d664e7fdae2e562baf23016a53943168af716e873"} Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.248026 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f6d895595-jv76b" event={"ID":"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba","Type":"ContainerStarted","Data":"20a20a4e69bdd787df49c91056089308888f8504fc31300ec9ca36a389aeb510"} Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.248144 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f6d895595-jv76b" event={"ID":"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba","Type":"ContainerStarted","Data":"60a3d290ffb9e90b0086750d7b0d89ef70348c9efbf1a8dea1b64dbf0d73f9de"} Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.248206 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f6d895595-jv76b" event={"ID":"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba","Type":"ContainerStarted","Data":"65e13bd44278107300790a93bc76ecae382f285e1385cee5ef8b18b0a1bbed0b"} Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.248288 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.248321 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.263603 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.483591131 podStartE2EDuration="10.263583255s" podCreationTimestamp="2025-10-05 07:13:12 +0000 UTC" firstStartedPulling="2025-10-05 07:13:13.023505669 +0000 UTC m=+1226.906132129" lastFinishedPulling="2025-10-05 07:13:21.803497783 +0000 UTC m=+1235.686124253" observedRunningTime="2025-10-05 07:13:22.260010601 +0000 UTC m=+1236.142637071" watchObservedRunningTime="2025-10-05 07:13:22.263583255 +0000 UTC m=+1236.146209715" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.294224 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6f6d895595-jv76b" podStartSLOduration=7.294198875 podStartE2EDuration="7.294198875s" podCreationTimestamp="2025-10-05 07:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:13:22.287540749 +0000 UTC m=+1236.170167199" watchObservedRunningTime="2025-10-05 07:13:22.294198875 +0000 UTC m=+1236.176825335" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.534165 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.672974 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-config-data\") pod \"6b527f48-ab8c-41e2-89b6-08e2456c687a\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.673060 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-sg-core-conf-yaml\") pod \"6b527f48-ab8c-41e2-89b6-08e2456c687a\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.673091 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv9ll\" (UniqueName: \"kubernetes.io/projected/6b527f48-ab8c-41e2-89b6-08e2456c687a-kube-api-access-cv9ll\") pod \"6b527f48-ab8c-41e2-89b6-08e2456c687a\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.673165 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-combined-ca-bundle\") pod \"6b527f48-ab8c-41e2-89b6-08e2456c687a\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.673243 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-scripts\") pod \"6b527f48-ab8c-41e2-89b6-08e2456c687a\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.673290 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-run-httpd\") pod \"6b527f48-ab8c-41e2-89b6-08e2456c687a\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.673649 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6b527f48-ab8c-41e2-89b6-08e2456c687a" (UID: "6b527f48-ab8c-41e2-89b6-08e2456c687a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.673763 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-log-httpd\") pod \"6b527f48-ab8c-41e2-89b6-08e2456c687a\" (UID: \"6b527f48-ab8c-41e2-89b6-08e2456c687a\") " Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.675917 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6b527f48-ab8c-41e2-89b6-08e2456c687a" (UID: "6b527f48-ab8c-41e2-89b6-08e2456c687a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.678219 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.678393 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b527f48-ab8c-41e2-89b6-08e2456c687a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.678660 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b527f48-ab8c-41e2-89b6-08e2456c687a-kube-api-access-cv9ll" (OuterVolumeSpecName: "kube-api-access-cv9ll") pod "6b527f48-ab8c-41e2-89b6-08e2456c687a" (UID: "6b527f48-ab8c-41e2-89b6-08e2456c687a"). InnerVolumeSpecName "kube-api-access-cv9ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.681145 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-scripts" (OuterVolumeSpecName: "scripts") pod "6b527f48-ab8c-41e2-89b6-08e2456c687a" (UID: "6b527f48-ab8c-41e2-89b6-08e2456c687a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.710092 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6b527f48-ab8c-41e2-89b6-08e2456c687a" (UID: "6b527f48-ab8c-41e2-89b6-08e2456c687a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.766773 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b527f48-ab8c-41e2-89b6-08e2456c687a" (UID: "6b527f48-ab8c-41e2-89b6-08e2456c687a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.779778 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.779807 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv9ll\" (UniqueName: \"kubernetes.io/projected/6b527f48-ab8c-41e2-89b6-08e2456c687a-kube-api-access-cv9ll\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.779819 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.779828 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.801602 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-config-data" (OuterVolumeSpecName: "config-data") pod "6b527f48-ab8c-41e2-89b6-08e2456c687a" (UID: "6b527f48-ab8c-41e2-89b6-08e2456c687a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:22 crc kubenswrapper[4935]: I1005 07:13:22.881801 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b527f48-ab8c-41e2-89b6-08e2456c687a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.260851 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerID="39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c" exitCode=0 Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.262057 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.264050 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerDied","Data":"39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c"} Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.264122 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b527f48-ab8c-41e2-89b6-08e2456c687a","Type":"ContainerDied","Data":"21bc9df07348db72d14c7b356af2f4f90e9f138012469e9f0bfc480eecb2253a"} Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.264145 4935 scope.go:117] "RemoveContainer" containerID="fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.290158 4935 scope.go:117] "RemoveContainer" containerID="b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.324408 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.341002 4935 scope.go:117] "RemoveContainer" containerID="39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.348083 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.372640 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:23 crc kubenswrapper[4935]: E1005 07:13:23.381113 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="ceilometer-notification-agent" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.381165 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="ceilometer-notification-agent" Oct 05 07:13:23 crc kubenswrapper[4935]: E1005 07:13:23.381201 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="sg-core" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.381210 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="sg-core" Oct 05 07:13:23 crc kubenswrapper[4935]: E1005 07:13:23.381237 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="ceilometer-central-agent" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.381245 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="ceilometer-central-agent" Oct 05 07:13:23 crc kubenswrapper[4935]: E1005 07:13:23.381271 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="proxy-httpd" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.381278 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="proxy-httpd" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.381562 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="ceilometer-central-agent" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.381584 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="proxy-httpd" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.381596 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="ceilometer-notification-agent" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.381612 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" containerName="sg-core" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.383694 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.390054 4935 scope.go:117] "RemoveContainer" containerID="835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.390289 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.390459 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.392322 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.461373 4935 scope.go:117] "RemoveContainer" containerID="fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e" Oct 05 07:13:23 crc kubenswrapper[4935]: E1005 07:13:23.461841 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e\": container with ID starting with fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e not found: ID does not exist" containerID="fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.461881 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e"} err="failed to get container status \"fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e\": rpc error: code = NotFound desc = could not find container \"fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e\": container with ID starting with fbc0565d33f91b615bb785dbc533dcfa1ac71d5f140b14a95e355da328c2d08e not found: ID does not exist" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.461916 4935 scope.go:117] "RemoveContainer" containerID="b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c" Oct 05 07:13:23 crc kubenswrapper[4935]: E1005 07:13:23.462254 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c\": container with ID starting with b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c not found: ID does not exist" containerID="b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.462275 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c"} err="failed to get container status \"b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c\": rpc error: code = NotFound desc = could not find container \"b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c\": container with ID starting with b7836070ce91af52312a5d8daf0058a3407bad2c7c18887d2991b14bf59ce77c not found: ID does not exist" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.462289 4935 scope.go:117] "RemoveContainer" containerID="39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c" Oct 05 07:13:23 crc kubenswrapper[4935]: E1005 07:13:23.462499 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c\": container with ID starting with 39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c not found: ID does not exist" containerID="39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.462517 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c"} err="failed to get container status \"39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c\": rpc error: code = NotFound desc = could not find container \"39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c\": container with ID starting with 39358ef70d002ad91f9718eff27d46e6f3c98d65f87c9ad35cb0ff3e1b2f280c not found: ID does not exist" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.462531 4935 scope.go:117] "RemoveContainer" containerID="835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69" Oct 05 07:13:23 crc kubenswrapper[4935]: E1005 07:13:23.462776 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69\": container with ID starting with 835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69 not found: ID does not exist" containerID="835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.462793 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69"} err="failed to get container status \"835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69\": rpc error: code = NotFound desc = could not find container \"835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69\": container with ID starting with 835d93ff42693000c4c024d2d357db35939a68f9431de5a40aeea169742dee69 not found: ID does not exist" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.500349 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-config-data\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.500443 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-scripts\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.500518 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-log-httpd\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.500574 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spqqz\" (UniqueName: \"kubernetes.io/projected/be1be8df-b209-475f-b3cd-132e411f05f9-kube-api-access-spqqz\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.500601 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.500624 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.500654 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-run-httpd\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.602363 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.602425 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-run-httpd\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.602473 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-config-data\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.602530 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-scripts\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.602560 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-log-httpd\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.602598 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spqqz\" (UniqueName: \"kubernetes.io/projected/be1be8df-b209-475f-b3cd-132e411f05f9-kube-api-access-spqqz\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.602618 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.603073 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-run-httpd\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.603423 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-log-httpd\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.606911 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-config-data\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.607488 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.607578 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-scripts\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.609354 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.622404 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spqqz\" (UniqueName: \"kubernetes.io/projected/be1be8df-b209-475f-b3cd-132e411f05f9-kube-api-access-spqqz\") pod \"ceilometer-0\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.681266 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-rnhsc"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.707823 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rnhsc" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.717617 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rnhsc"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.719863 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.764944 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.793172 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-wfm9k"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.794794 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wfm9k" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.809882 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwh54\" (UniqueName: \"kubernetes.io/projected/a3f78262-5eb9-4a4d-9a93-2333f428b032-kube-api-access-pwh54\") pod \"nova-api-db-create-rnhsc\" (UID: \"a3f78262-5eb9-4a4d-9a93-2333f428b032\") " pod="openstack/nova-api-db-create-rnhsc" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.811924 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wfm9k"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.911451 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-4wf8f"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.913074 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4wf8f" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.919003 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4wf8f"] Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.920490 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk2v6\" (UniqueName: \"kubernetes.io/projected/2967548c-91b1-43de-9c7d-53b462b09478-kube-api-access-gk2v6\") pod \"nova-cell0-db-create-wfm9k\" (UID: \"2967548c-91b1-43de-9c7d-53b462b09478\") " pod="openstack/nova-cell0-db-create-wfm9k" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.920561 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwh54\" (UniqueName: \"kubernetes.io/projected/a3f78262-5eb9-4a4d-9a93-2333f428b032-kube-api-access-pwh54\") pod \"nova-api-db-create-rnhsc\" (UID: \"a3f78262-5eb9-4a4d-9a93-2333f428b032\") " pod="openstack/nova-api-db-create-rnhsc" Oct 05 07:13:23 crc kubenswrapper[4935]: I1005 07:13:23.967718 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwh54\" (UniqueName: \"kubernetes.io/projected/a3f78262-5eb9-4a4d-9a93-2333f428b032-kube-api-access-pwh54\") pod \"nova-api-db-create-rnhsc\" (UID: \"a3f78262-5eb9-4a4d-9a93-2333f428b032\") " pod="openstack/nova-api-db-create-rnhsc" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.022507 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgfpn\" (UniqueName: \"kubernetes.io/projected/9d26dccc-7629-42a9-9597-600f4750c11b-kube-api-access-rgfpn\") pod \"nova-cell1-db-create-4wf8f\" (UID: \"9d26dccc-7629-42a9-9597-600f4750c11b\") " pod="openstack/nova-cell1-db-create-4wf8f" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.022629 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk2v6\" (UniqueName: \"kubernetes.io/projected/2967548c-91b1-43de-9c7d-53b462b09478-kube-api-access-gk2v6\") pod \"nova-cell0-db-create-wfm9k\" (UID: \"2967548c-91b1-43de-9c7d-53b462b09478\") " pod="openstack/nova-cell0-db-create-wfm9k" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.053656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk2v6\" (UniqueName: \"kubernetes.io/projected/2967548c-91b1-43de-9c7d-53b462b09478-kube-api-access-gk2v6\") pod \"nova-cell0-db-create-wfm9k\" (UID: \"2967548c-91b1-43de-9c7d-53b462b09478\") " pod="openstack/nova-cell0-db-create-wfm9k" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.124507 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgfpn\" (UniqueName: \"kubernetes.io/projected/9d26dccc-7629-42a9-9597-600f4750c11b-kube-api-access-rgfpn\") pod \"nova-cell1-db-create-4wf8f\" (UID: \"9d26dccc-7629-42a9-9597-600f4750c11b\") " pod="openstack/nova-cell1-db-create-4wf8f" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.140621 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgfpn\" (UniqueName: \"kubernetes.io/projected/9d26dccc-7629-42a9-9597-600f4750c11b-kube-api-access-rgfpn\") pod \"nova-cell1-db-create-4wf8f\" (UID: \"9d26dccc-7629-42a9-9597-600f4750c11b\") " pod="openstack/nova-cell1-db-create-4wf8f" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.145705 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rnhsc" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.161706 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wfm9k" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.310756 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4wf8f" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.310802 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.649232 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wfm9k"] Oct 05 07:13:24 crc kubenswrapper[4935]: W1005 07:13:24.708688 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3f78262_5eb9_4a4d_9a93_2333f428b032.slice/crio-15214f5ccfe432647df9381e49174384a6b3678e281e07b746d8a3640c654495 WatchSource:0}: Error finding container 15214f5ccfe432647df9381e49174384a6b3678e281e07b746d8a3640c654495: Status 404 returned error can't find the container with id 15214f5ccfe432647df9381e49174384a6b3678e281e07b746d8a3640c654495 Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.712927 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rnhsc"] Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.791710 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b527f48-ab8c-41e2-89b6-08e2456c687a" path="/var/lib/kubelet/pods/6b527f48-ab8c-41e2-89b6-08e2456c687a/volumes" Oct 05 07:13:24 crc kubenswrapper[4935]: I1005 07:13:24.821165 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4wf8f"] Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.283104 4935 generic.go:334] "Generic (PLEG): container finished" podID="a3f78262-5eb9-4a4d-9a93-2333f428b032" containerID="e86b66c8e04febdd2bd5944c182e0ff7d4dbf92d9fc614f144b437495b40e51c" exitCode=0 Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.283196 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rnhsc" event={"ID":"a3f78262-5eb9-4a4d-9a93-2333f428b032","Type":"ContainerDied","Data":"e86b66c8e04febdd2bd5944c182e0ff7d4dbf92d9fc614f144b437495b40e51c"} Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.283836 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rnhsc" event={"ID":"a3f78262-5eb9-4a4d-9a93-2333f428b032","Type":"ContainerStarted","Data":"15214f5ccfe432647df9381e49174384a6b3678e281e07b746d8a3640c654495"} Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.286531 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerStarted","Data":"3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce"} Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.286578 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerStarted","Data":"a2975d39fd26135e00c850a1125d7caba76f2465c43610229877e6b8b0c50296"} Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.288405 4935 generic.go:334] "Generic (PLEG): container finished" podID="2967548c-91b1-43de-9c7d-53b462b09478" containerID="886247bb7fb5027803fe13e7b67575d77b9ea5c583be3ae7be247796ac3beed7" exitCode=0 Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.288476 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wfm9k" event={"ID":"2967548c-91b1-43de-9c7d-53b462b09478","Type":"ContainerDied","Data":"886247bb7fb5027803fe13e7b67575d77b9ea5c583be3ae7be247796ac3beed7"} Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.288508 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wfm9k" event={"ID":"2967548c-91b1-43de-9c7d-53b462b09478","Type":"ContainerStarted","Data":"b13460c8ee0f5ae2da81f8bcc6fdc6f5939c355d95d8f851dbad34d76ff67c1e"} Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.290256 4935 generic.go:334] "Generic (PLEG): container finished" podID="9d26dccc-7629-42a9-9597-600f4750c11b" containerID="e8f8d95dd578c7c0d462ac9775a7a22a446295ed6782e88a6b08a05a1619462d" exitCode=0 Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.290295 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4wf8f" event={"ID":"9d26dccc-7629-42a9-9597-600f4750c11b","Type":"ContainerDied","Data":"e8f8d95dd578c7c0d462ac9775a7a22a446295ed6782e88a6b08a05a1619462d"} Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.290315 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4wf8f" event={"ID":"9d26dccc-7629-42a9-9597-600f4750c11b","Type":"ContainerStarted","Data":"eeb321b983e2293b0912c6e5fb74ac58c512cb82ddb0dde62ea9b162607b4388"} Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.472825 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.473108 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="980fa02a-09b0-42cc-8028-51b588427002" containerName="glance-log" containerID="cri-o://ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362" gracePeriod=30 Oct 05 07:13:25 crc kubenswrapper[4935]: I1005 07:13:25.473192 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="980fa02a-09b0-42cc-8028-51b588427002" containerName="glance-httpd" containerID="cri-o://9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a" gracePeriod=30 Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.307660 4935 generic.go:334] "Generic (PLEG): container finished" podID="980fa02a-09b0-42cc-8028-51b588427002" containerID="ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362" exitCode=143 Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.307853 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"980fa02a-09b0-42cc-8028-51b588427002","Type":"ContainerDied","Data":"ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362"} Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.312136 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerStarted","Data":"e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256"} Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.315008 4935 generic.go:334] "Generic (PLEG): container finished" podID="e94e21e9-e1a5-406b-9235-35165cded309" containerID="b4761f2806893419fa049bdae0c7dd84eb9621a5db9d1c6aa5094759cbd94a58" exitCode=137 Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.315066 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e94e21e9-e1a5-406b-9235-35165cded309","Type":"ContainerDied","Data":"b4761f2806893419fa049bdae0c7dd84eb9621a5db9d1c6aa5094759cbd94a58"} Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.315348 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e94e21e9-e1a5-406b-9235-35165cded309","Type":"ContainerDied","Data":"f47cbeb7226c7fb1e8f46f5b604659173cebbbde55bea4e77c469cd787378f1a"} Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.315503 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f47cbeb7226c7fb1e8f46f5b604659173cebbbde55bea4e77c469cd787378f1a" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.331022 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.510765 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-scripts\") pod \"e94e21e9-e1a5-406b-9235-35165cded309\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.510850 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7tnl\" (UniqueName: \"kubernetes.io/projected/e94e21e9-e1a5-406b-9235-35165cded309-kube-api-access-w7tnl\") pod \"e94e21e9-e1a5-406b-9235-35165cded309\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.510905 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e94e21e9-e1a5-406b-9235-35165cded309-logs\") pod \"e94e21e9-e1a5-406b-9235-35165cded309\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.510931 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e94e21e9-e1a5-406b-9235-35165cded309-etc-machine-id\") pod \"e94e21e9-e1a5-406b-9235-35165cded309\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.510950 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-combined-ca-bundle\") pod \"e94e21e9-e1a5-406b-9235-35165cded309\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.511043 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data\") pod \"e94e21e9-e1a5-406b-9235-35165cded309\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.511244 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data-custom\") pod \"e94e21e9-e1a5-406b-9235-35165cded309\" (UID: \"e94e21e9-e1a5-406b-9235-35165cded309\") " Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.511788 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e94e21e9-e1a5-406b-9235-35165cded309-logs" (OuterVolumeSpecName: "logs") pod "e94e21e9-e1a5-406b-9235-35165cded309" (UID: "e94e21e9-e1a5-406b-9235-35165cded309"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.516718 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-scripts" (OuterVolumeSpecName: "scripts") pod "e94e21e9-e1a5-406b-9235-35165cded309" (UID: "e94e21e9-e1a5-406b-9235-35165cded309"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.517666 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e94e21e9-e1a5-406b-9235-35165cded309" (UID: "e94e21e9-e1a5-406b-9235-35165cded309"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.517733 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e94e21e9-e1a5-406b-9235-35165cded309-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e94e21e9-e1a5-406b-9235-35165cded309" (UID: "e94e21e9-e1a5-406b-9235-35165cded309"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.523318 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94e21e9-e1a5-406b-9235-35165cded309-kube-api-access-w7tnl" (OuterVolumeSpecName: "kube-api-access-w7tnl") pod "e94e21e9-e1a5-406b-9235-35165cded309" (UID: "e94e21e9-e1a5-406b-9235-35165cded309"). InnerVolumeSpecName "kube-api-access-w7tnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.540492 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e94e21e9-e1a5-406b-9235-35165cded309" (UID: "e94e21e9-e1a5-406b-9235-35165cded309"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.574937 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data" (OuterVolumeSpecName: "config-data") pod "e94e21e9-e1a5-406b-9235-35165cded309" (UID: "e94e21e9-e1a5-406b-9235-35165cded309"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.614787 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.614832 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.614847 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.614859 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7tnl\" (UniqueName: \"kubernetes.io/projected/e94e21e9-e1a5-406b-9235-35165cded309-kube-api-access-w7tnl\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.614871 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e94e21e9-e1a5-406b-9235-35165cded309-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.614879 4935 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e94e21e9-e1a5-406b-9235-35165cded309-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.614905 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94e21e9-e1a5-406b-9235-35165cded309-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.699611 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rnhsc" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.715537 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwh54\" (UniqueName: \"kubernetes.io/projected/a3f78262-5eb9-4a4d-9a93-2333f428b032-kube-api-access-pwh54\") pod \"a3f78262-5eb9-4a4d-9a93-2333f428b032\" (UID: \"a3f78262-5eb9-4a4d-9a93-2333f428b032\") " Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.730791 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f78262-5eb9-4a4d-9a93-2333f428b032-kube-api-access-pwh54" (OuterVolumeSpecName: "kube-api-access-pwh54") pod "a3f78262-5eb9-4a4d-9a93-2333f428b032" (UID: "a3f78262-5eb9-4a4d-9a93-2333f428b032"). InnerVolumeSpecName "kube-api-access-pwh54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.826734 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwh54\" (UniqueName: \"kubernetes.io/projected/a3f78262-5eb9-4a4d-9a93-2333f428b032-kube-api-access-pwh54\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.951503 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4wf8f" Oct 05 07:13:26 crc kubenswrapper[4935]: I1005 07:13:26.957563 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wfm9k" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.132568 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgfpn\" (UniqueName: \"kubernetes.io/projected/9d26dccc-7629-42a9-9597-600f4750c11b-kube-api-access-rgfpn\") pod \"9d26dccc-7629-42a9-9597-600f4750c11b\" (UID: \"9d26dccc-7629-42a9-9597-600f4750c11b\") " Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.132694 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk2v6\" (UniqueName: \"kubernetes.io/projected/2967548c-91b1-43de-9c7d-53b462b09478-kube-api-access-gk2v6\") pod \"2967548c-91b1-43de-9c7d-53b462b09478\" (UID: \"2967548c-91b1-43de-9c7d-53b462b09478\") " Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.136192 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2967548c-91b1-43de-9c7d-53b462b09478-kube-api-access-gk2v6" (OuterVolumeSpecName: "kube-api-access-gk2v6") pod "2967548c-91b1-43de-9c7d-53b462b09478" (UID: "2967548c-91b1-43de-9c7d-53b462b09478"). InnerVolumeSpecName "kube-api-access-gk2v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.139042 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d26dccc-7629-42a9-9597-600f4750c11b-kube-api-access-rgfpn" (OuterVolumeSpecName: "kube-api-access-rgfpn") pod "9d26dccc-7629-42a9-9597-600f4750c11b" (UID: "9d26dccc-7629-42a9-9597-600f4750c11b"). InnerVolumeSpecName "kube-api-access-rgfpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.234591 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgfpn\" (UniqueName: \"kubernetes.io/projected/9d26dccc-7629-42a9-9597-600f4750c11b-kube-api-access-rgfpn\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.234620 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk2v6\" (UniqueName: \"kubernetes.io/projected/2967548c-91b1-43de-9c7d-53b462b09478-kube-api-access-gk2v6\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.325246 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rnhsc" event={"ID":"a3f78262-5eb9-4a4d-9a93-2333f428b032","Type":"ContainerDied","Data":"15214f5ccfe432647df9381e49174384a6b3678e281e07b746d8a3640c654495"} Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.325631 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15214f5ccfe432647df9381e49174384a6b3678e281e07b746d8a3640c654495" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.325711 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rnhsc" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.328588 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerStarted","Data":"828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4"} Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.331952 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wfm9k" event={"ID":"2967548c-91b1-43de-9c7d-53b462b09478","Type":"ContainerDied","Data":"b13460c8ee0f5ae2da81f8bcc6fdc6f5939c355d95d8f851dbad34d76ff67c1e"} Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.332004 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b13460c8ee0f5ae2da81f8bcc6fdc6f5939c355d95d8f851dbad34d76ff67c1e" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.331972 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wfm9k" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.343278 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.344169 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4wf8f" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.345320 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4wf8f" event={"ID":"9d26dccc-7629-42a9-9597-600f4750c11b","Type":"ContainerDied","Data":"eeb321b983e2293b0912c6e5fb74ac58c512cb82ddb0dde62ea9b162607b4388"} Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.345380 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeb321b983e2293b0912c6e5fb74ac58c512cb82ddb0dde62ea9b162607b4388" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.370866 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.376935 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.403766 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:13:27 crc kubenswrapper[4935]: E1005 07:13:27.404200 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2967548c-91b1-43de-9c7d-53b462b09478" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404219 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2967548c-91b1-43de-9c7d-53b462b09478" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: E1005 07:13:27.404244 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f78262-5eb9-4a4d-9a93-2333f428b032" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404250 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f78262-5eb9-4a4d-9a93-2333f428b032" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: E1005 07:13:27.404262 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d26dccc-7629-42a9-9597-600f4750c11b" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404269 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d26dccc-7629-42a9-9597-600f4750c11b" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: E1005 07:13:27.404276 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94e21e9-e1a5-406b-9235-35165cded309" containerName="cinder-api" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404282 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94e21e9-e1a5-406b-9235-35165cded309" containerName="cinder-api" Oct 05 07:13:27 crc kubenswrapper[4935]: E1005 07:13:27.404302 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94e21e9-e1a5-406b-9235-35165cded309" containerName="cinder-api-log" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404308 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94e21e9-e1a5-406b-9235-35165cded309" containerName="cinder-api-log" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404490 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e94e21e9-e1a5-406b-9235-35165cded309" containerName="cinder-api-log" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404501 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f78262-5eb9-4a4d-9a93-2333f428b032" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404522 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e94e21e9-e1a5-406b-9235-35165cded309" containerName="cinder-api" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404533 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d26dccc-7629-42a9-9597-600f4750c11b" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.404541 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2967548c-91b1-43de-9c7d-53b462b09478" containerName="mariadb-database-create" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.405490 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.408683 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.408869 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.410780 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.412188 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.540730 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data-custom\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.540779 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjx6t\" (UniqueName: \"kubernetes.io/projected/7c7acc7a-797f-46cd-b298-1f28fd2951d6-kube-api-access-jjx6t\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.540828 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c7acc7a-797f-46cd-b298-1f28fd2951d6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.541008 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.541117 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.541262 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.541321 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c7acc7a-797f-46cd-b298-1f28fd2951d6-logs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.541405 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.541439 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-scripts\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.643784 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.643879 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.643966 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.644013 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c7acc7a-797f-46cd-b298-1f28fd2951d6-logs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.644098 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.644125 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-scripts\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.644165 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data-custom\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.644195 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjx6t\" (UniqueName: \"kubernetes.io/projected/7c7acc7a-797f-46cd-b298-1f28fd2951d6-kube-api-access-jjx6t\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.644250 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c7acc7a-797f-46cd-b298-1f28fd2951d6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.644372 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c7acc7a-797f-46cd-b298-1f28fd2951d6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.645090 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c7acc7a-797f-46cd-b298-1f28fd2951d6-logs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.648195 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.648609 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-scripts\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.649172 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data-custom\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.649361 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.650642 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.661486 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.666351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjx6t\" (UniqueName: \"kubernetes.io/projected/7c7acc7a-797f-46cd-b298-1f28fd2951d6-kube-api-access-jjx6t\") pod \"cinder-api-0\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " pod="openstack/cinder-api-0" Oct 05 07:13:27 crc kubenswrapper[4935]: I1005 07:13:27.721275 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.160200 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.352525 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7c7acc7a-797f-46cd-b298-1f28fd2951d6","Type":"ContainerStarted","Data":"7d70d45c48415ff1f6889a218122105e1192896f2ffe16de0dff780b19f7a4c5"} Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.355285 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerStarted","Data":"f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746"} Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.355485 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="ceilometer-central-agent" containerID="cri-o://3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce" gracePeriod=30 Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.355835 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.356179 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="proxy-httpd" containerID="cri-o://f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746" gracePeriod=30 Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.356247 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="sg-core" containerID="cri-o://828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4" gracePeriod=30 Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.356302 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="ceilometer-notification-agent" containerID="cri-o://e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256" gracePeriod=30 Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.383939 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.284678027 podStartE2EDuration="5.379821087s" podCreationTimestamp="2025-10-05 07:13:23 +0000 UTC" firstStartedPulling="2025-10-05 07:13:24.335927495 +0000 UTC m=+1238.218553955" lastFinishedPulling="2025-10-05 07:13:27.431070555 +0000 UTC m=+1241.313697015" observedRunningTime="2025-10-05 07:13:28.373706325 +0000 UTC m=+1242.256332785" watchObservedRunningTime="2025-10-05 07:13:28.379821087 +0000 UTC m=+1242.262447547" Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.586722 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.587574 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerName="glance-log" containerID="cri-o://9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4" gracePeriod=30 Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.588952 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerName="glance-httpd" containerID="cri-o://9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c" gracePeriod=30 Oct 05 07:13:28 crc kubenswrapper[4935]: I1005 07:13:28.813797 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94e21e9-e1a5-406b-9235-35165cded309" path="/var/lib/kubelet/pods/e94e21e9-e1a5-406b-9235-35165cded309/volumes" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.124850 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.287649 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-scripts\") pod \"980fa02a-09b0-42cc-8028-51b588427002\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.287732 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-config-data\") pod \"980fa02a-09b0-42cc-8028-51b588427002\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.287794 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-logs\") pod \"980fa02a-09b0-42cc-8028-51b588427002\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.287826 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-httpd-run\") pod \"980fa02a-09b0-42cc-8028-51b588427002\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.287877 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7jcq\" (UniqueName: \"kubernetes.io/projected/980fa02a-09b0-42cc-8028-51b588427002-kube-api-access-b7jcq\") pod \"980fa02a-09b0-42cc-8028-51b588427002\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.287918 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-internal-tls-certs\") pod \"980fa02a-09b0-42cc-8028-51b588427002\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.287958 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-combined-ca-bundle\") pod \"980fa02a-09b0-42cc-8028-51b588427002\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.288005 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"980fa02a-09b0-42cc-8028-51b588427002\" (UID: \"980fa02a-09b0-42cc-8028-51b588427002\") " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.288407 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-logs" (OuterVolumeSpecName: "logs") pod "980fa02a-09b0-42cc-8028-51b588427002" (UID: "980fa02a-09b0-42cc-8028-51b588427002"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.289185 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "980fa02a-09b0-42cc-8028-51b588427002" (UID: "980fa02a-09b0-42cc-8028-51b588427002"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.292145 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "980fa02a-09b0-42cc-8028-51b588427002" (UID: "980fa02a-09b0-42cc-8028-51b588427002"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.292375 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-scripts" (OuterVolumeSpecName: "scripts") pod "980fa02a-09b0-42cc-8028-51b588427002" (UID: "980fa02a-09b0-42cc-8028-51b588427002"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.295316 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/980fa02a-09b0-42cc-8028-51b588427002-kube-api-access-b7jcq" (OuterVolumeSpecName: "kube-api-access-b7jcq") pod "980fa02a-09b0-42cc-8028-51b588427002" (UID: "980fa02a-09b0-42cc-8028-51b588427002"). InnerVolumeSpecName "kube-api-access-b7jcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.339880 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "980fa02a-09b0-42cc-8028-51b588427002" (UID: "980fa02a-09b0-42cc-8028-51b588427002"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.353103 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-config-data" (OuterVolumeSpecName: "config-data") pod "980fa02a-09b0-42cc-8028-51b588427002" (UID: "980fa02a-09b0-42cc-8028-51b588427002"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.364796 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "980fa02a-09b0-42cc-8028-51b588427002" (UID: "980fa02a-09b0-42cc-8028-51b588427002"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.372503 4935 generic.go:334] "Generic (PLEG): container finished" podID="980fa02a-09b0-42cc-8028-51b588427002" containerID="9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a" exitCode=0 Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.372568 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.372705 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"980fa02a-09b0-42cc-8028-51b588427002","Type":"ContainerDied","Data":"9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a"} Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.372764 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"980fa02a-09b0-42cc-8028-51b588427002","Type":"ContainerDied","Data":"19f1e4a6406ea7af77cb699b64a82111391d5c4cdb9f6da517169509001317e7"} Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.372795 4935 scope.go:117] "RemoveContainer" containerID="9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.377576 4935 generic.go:334] "Generic (PLEG): container finished" podID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerID="9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4" exitCode=143 Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.377651 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b885bd46-8ccc-47d6-8622-c4175e2267e5","Type":"ContainerDied","Data":"9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4"} Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.381703 4935 generic.go:334] "Generic (PLEG): container finished" podID="be1be8df-b209-475f-b3cd-132e411f05f9" containerID="f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746" exitCode=0 Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.381747 4935 generic.go:334] "Generic (PLEG): container finished" podID="be1be8df-b209-475f-b3cd-132e411f05f9" containerID="828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4" exitCode=2 Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.381755 4935 generic.go:334] "Generic (PLEG): container finished" podID="be1be8df-b209-475f-b3cd-132e411f05f9" containerID="e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256" exitCode=0 Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.381813 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerDied","Data":"f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746"} Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.381844 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerDied","Data":"828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4"} Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.381854 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerDied","Data":"e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256"} Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.383871 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7c7acc7a-797f-46cd-b298-1f28fd2951d6","Type":"ContainerStarted","Data":"afd975882d932c2ed56cbd1069d50977a2d2abf98e1305a15671c213fc965de8"} Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.389656 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.390030 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.390042 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.390051 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/980fa02a-09b0-42cc-8028-51b588427002-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.390061 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7jcq\" (UniqueName: \"kubernetes.io/projected/980fa02a-09b0-42cc-8028-51b588427002-kube-api-access-b7jcq\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.390071 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.390079 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/980fa02a-09b0-42cc-8028-51b588427002-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.390109 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.409503 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.419532 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.428765 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.434266 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:13:29 crc kubenswrapper[4935]: E1005 07:13:29.434781 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980fa02a-09b0-42cc-8028-51b588427002" containerName="glance-log" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.434806 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="980fa02a-09b0-42cc-8028-51b588427002" containerName="glance-log" Oct 05 07:13:29 crc kubenswrapper[4935]: E1005 07:13:29.434873 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980fa02a-09b0-42cc-8028-51b588427002" containerName="glance-httpd" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.434881 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="980fa02a-09b0-42cc-8028-51b588427002" containerName="glance-httpd" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.435108 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="980fa02a-09b0-42cc-8028-51b588427002" containerName="glance-httpd" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.435133 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="980fa02a-09b0-42cc-8028-51b588427002" containerName="glance-log" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.436868 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.442401 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.442721 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.446902 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.451037 4935 scope.go:117] "RemoveContainer" containerID="ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.494257 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.537193 4935 scope.go:117] "RemoveContainer" containerID="9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a" Oct 05 07:13:29 crc kubenswrapper[4935]: E1005 07:13:29.537680 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a\": container with ID starting with 9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a not found: ID does not exist" containerID="9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.537707 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a"} err="failed to get container status \"9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a\": rpc error: code = NotFound desc = could not find container \"9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a\": container with ID starting with 9a2c2df444138b4fdaafd007bde3438c1ffa126470f461ea52af7fc7fb84567a not found: ID does not exist" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.537728 4935 scope.go:117] "RemoveContainer" containerID="ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362" Oct 05 07:13:29 crc kubenswrapper[4935]: E1005 07:13:29.537956 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362\": container with ID starting with ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362 not found: ID does not exist" containerID="ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.537978 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362"} err="failed to get container status \"ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362\": rpc error: code = NotFound desc = could not find container \"ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362\": container with ID starting with ce8f1b6404061cab9b8ba009b373c9b88794ec444bce087eda3a9d492516c362 not found: ID does not exist" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.597972 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.598226 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.598345 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.598580 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.598704 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx9rd\" (UniqueName: \"kubernetes.io/projected/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-kube-api-access-cx9rd\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.598790 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.598817 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.598987 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701006 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701121 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701160 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx9rd\" (UniqueName: \"kubernetes.io/projected/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-kube-api-access-cx9rd\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701189 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701207 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701238 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701292 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701327 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701467 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.701858 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.706282 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.706735 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.708793 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.710398 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.722789 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.725155 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx9rd\" (UniqueName: \"kubernetes.io/projected/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-kube-api-access-cx9rd\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.740387 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:13:29 crc kubenswrapper[4935]: I1005 07:13:29.775325 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:30 crc kubenswrapper[4935]: W1005 07:13:30.351444 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74e20071_1ff2_4fae_ae8c_beb9dbce35ce.slice/crio-029b36e747b3b514a2e5123e18003f92080faaf01ed4d2dc433120d5adcbe579 WatchSource:0}: Error finding container 029b36e747b3b514a2e5123e18003f92080faaf01ed4d2dc433120d5adcbe579: Status 404 returned error can't find the container with id 029b36e747b3b514a2e5123e18003f92080faaf01ed4d2dc433120d5adcbe579 Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.352131 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.395604 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74e20071-1ff2-4fae-ae8c-beb9dbce35ce","Type":"ContainerStarted","Data":"029b36e747b3b514a2e5123e18003f92080faaf01ed4d2dc433120d5adcbe579"} Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.401299 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7c7acc7a-797f-46cd-b298-1f28fd2951d6","Type":"ContainerStarted","Data":"7378ebd5875058e6928f9504c6ce673a929f0b593e3c881ba94728592aaa6422"} Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.401674 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.429627 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.429594818 podStartE2EDuration="3.429594818s" podCreationTimestamp="2025-10-05 07:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:13:30.421736901 +0000 UTC m=+1244.304363391" watchObservedRunningTime="2025-10-05 07:13:30.429594818 +0000 UTC m=+1244.312221298" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.789430 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="980fa02a-09b0-42cc-8028-51b588427002" path="/var/lib/kubelet/pods/980fa02a-09b0-42cc-8028-51b588427002/volumes" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.826174 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.843565 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-sg-core-conf-yaml\") pod \"be1be8df-b209-475f-b3cd-132e411f05f9\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.843607 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spqqz\" (UniqueName: \"kubernetes.io/projected/be1be8df-b209-475f-b3cd-132e411f05f9-kube-api-access-spqqz\") pod \"be1be8df-b209-475f-b3cd-132e411f05f9\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.843669 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-log-httpd\") pod \"be1be8df-b209-475f-b3cd-132e411f05f9\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.843699 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-run-httpd\") pod \"be1be8df-b209-475f-b3cd-132e411f05f9\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.843718 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-scripts\") pod \"be1be8df-b209-475f-b3cd-132e411f05f9\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.843777 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-config-data\") pod \"be1be8df-b209-475f-b3cd-132e411f05f9\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.843804 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-combined-ca-bundle\") pod \"be1be8df-b209-475f-b3cd-132e411f05f9\" (UID: \"be1be8df-b209-475f-b3cd-132e411f05f9\") " Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.855366 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "be1be8df-b209-475f-b3cd-132e411f05f9" (UID: "be1be8df-b209-475f-b3cd-132e411f05f9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.855646 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "be1be8df-b209-475f-b3cd-132e411f05f9" (UID: "be1be8df-b209-475f-b3cd-132e411f05f9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.857394 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be1be8df-b209-475f-b3cd-132e411f05f9-kube-api-access-spqqz" (OuterVolumeSpecName: "kube-api-access-spqqz") pod "be1be8df-b209-475f-b3cd-132e411f05f9" (UID: "be1be8df-b209-475f-b3cd-132e411f05f9"). InnerVolumeSpecName "kube-api-access-spqqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.882390 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-scripts" (OuterVolumeSpecName: "scripts") pod "be1be8df-b209-475f-b3cd-132e411f05f9" (UID: "be1be8df-b209-475f-b3cd-132e411f05f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.888068 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "be1be8df-b209-475f-b3cd-132e411f05f9" (UID: "be1be8df-b209-475f-b3cd-132e411f05f9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.946795 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.947245 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spqqz\" (UniqueName: \"kubernetes.io/projected/be1be8df-b209-475f-b3cd-132e411f05f9-kube-api-access-spqqz\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.947368 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.947481 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be1be8df-b209-475f-b3cd-132e411f05f9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.947583 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.971050 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be1be8df-b209-475f-b3cd-132e411f05f9" (UID: "be1be8df-b209-475f-b3cd-132e411f05f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:30 crc kubenswrapper[4935]: I1005 07:13:30.991390 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-config-data" (OuterVolumeSpecName: "config-data") pod "be1be8df-b209-475f-b3cd-132e411f05f9" (UID: "be1be8df-b209-475f-b3cd-132e411f05f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.048218 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.048456 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1be8df-b209-475f-b3cd-132e411f05f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.251210 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.255617 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.419093 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74e20071-1ff2-4fae-ae8c-beb9dbce35ce","Type":"ContainerStarted","Data":"af59bf08d75bada41121a63c5849541082f4196ebb916ac86beff7f230324a9b"} Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.431807 4935 generic.go:334] "Generic (PLEG): container finished" podID="be1be8df-b209-475f-b3cd-132e411f05f9" containerID="3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce" exitCode=0 Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.432790 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.440028 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerDied","Data":"3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce"} Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.440175 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be1be8df-b209-475f-b3cd-132e411f05f9","Type":"ContainerDied","Data":"a2975d39fd26135e00c850a1125d7caba76f2465c43610229877e6b8b0c50296"} Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.440228 4935 scope.go:117] "RemoveContainer" containerID="f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.481306 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.490813 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516247 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:31 crc kubenswrapper[4935]: E1005 07:13:31.516654 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="proxy-httpd" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516671 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="proxy-httpd" Oct 05 07:13:31 crc kubenswrapper[4935]: E1005 07:13:31.516689 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="sg-core" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516697 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="sg-core" Oct 05 07:13:31 crc kubenswrapper[4935]: E1005 07:13:31.516731 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="ceilometer-notification-agent" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516738 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="ceilometer-notification-agent" Oct 05 07:13:31 crc kubenswrapper[4935]: E1005 07:13:31.516749 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="ceilometer-central-agent" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516755 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="ceilometer-central-agent" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516914 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="ceilometer-notification-agent" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516935 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="sg-core" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516946 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="proxy-httpd" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.516958 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" containerName="ceilometer-central-agent" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.518557 4935 scope.go:117] "RemoveContainer" containerID="828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.520443 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.524343 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.525058 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.533800 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.549964 4935 scope.go:117] "RemoveContainer" containerID="e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.562765 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-log-httpd\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.562822 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-config-data\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.562845 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-run-httpd\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.562937 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b5mp\" (UniqueName: \"kubernetes.io/projected/5530001a-4323-49d7-adc3-eaa49ecb0963-kube-api-access-7b5mp\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.562956 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.562971 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-scripts\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.563042 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.580519 4935 scope.go:117] "RemoveContainer" containerID="3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.598609 4935 scope.go:117] "RemoveContainer" containerID="f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746" Oct 05 07:13:31 crc kubenswrapper[4935]: E1005 07:13:31.598993 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746\": container with ID starting with f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746 not found: ID does not exist" containerID="f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.599022 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746"} err="failed to get container status \"f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746\": rpc error: code = NotFound desc = could not find container \"f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746\": container with ID starting with f2a378985b07e8238efaf6570aede7b8cd8289e8c0fba6079d375e292edbe746 not found: ID does not exist" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.599043 4935 scope.go:117] "RemoveContainer" containerID="828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4" Oct 05 07:13:31 crc kubenswrapper[4935]: E1005 07:13:31.600215 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4\": container with ID starting with 828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4 not found: ID does not exist" containerID="828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.600277 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4"} err="failed to get container status \"828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4\": rpc error: code = NotFound desc = could not find container \"828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4\": container with ID starting with 828a296319cb5cb868e9526b8919810a938210a085e5417bf45f415aeca613e4 not found: ID does not exist" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.600305 4935 scope.go:117] "RemoveContainer" containerID="e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256" Oct 05 07:13:31 crc kubenswrapper[4935]: E1005 07:13:31.600683 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256\": container with ID starting with e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256 not found: ID does not exist" containerID="e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.600730 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256"} err="failed to get container status \"e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256\": rpc error: code = NotFound desc = could not find container \"e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256\": container with ID starting with e1ab33f16819e674b7e3300b5d5f3baefb2cc5b35ce45602c09731c6f1ebf256 not found: ID does not exist" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.600763 4935 scope.go:117] "RemoveContainer" containerID="3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce" Oct 05 07:13:31 crc kubenswrapper[4935]: E1005 07:13:31.601017 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce\": container with ID starting with 3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce not found: ID does not exist" containerID="3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.601046 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce"} err="failed to get container status \"3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce\": rpc error: code = NotFound desc = could not find container \"3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce\": container with ID starting with 3af4107aafb24882a003dd4d256aa4fe2df1bf25c56b595d7c7d0a16729831ce not found: ID does not exist" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.664237 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.664312 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-log-httpd\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.664344 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-config-data\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.664363 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-run-httpd\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.664430 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b5mp\" (UniqueName: \"kubernetes.io/projected/5530001a-4323-49d7-adc3-eaa49ecb0963-kube-api-access-7b5mp\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.664447 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.664462 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-scripts\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.665658 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-log-httpd\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.666044 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-run-httpd\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.670482 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.670550 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.671063 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-scripts\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.671199 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-config-data\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.682703 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b5mp\" (UniqueName: \"kubernetes.io/projected/5530001a-4323-49d7-adc3-eaa49ecb0963-kube-api-access-7b5mp\") pod \"ceilometer-0\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " pod="openstack/ceilometer-0" Oct 05 07:13:31 crc kubenswrapper[4935]: I1005 07:13:31.843691 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.410298 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.444717 4935 generic.go:334] "Generic (PLEG): container finished" podID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerID="9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c" exitCode=0 Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.444826 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b885bd46-8ccc-47d6-8622-c4175e2267e5","Type":"ContainerDied","Data":"9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c"} Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.444860 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b885bd46-8ccc-47d6-8622-c4175e2267e5","Type":"ContainerDied","Data":"7637814c072e95a7f542fd6a90d7331c8a2dfdb0f2b369847b4007abe680b762"} Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.444883 4935 scope.go:117] "RemoveContainer" containerID="9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.445045 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.457726 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74e20071-1ff2-4fae-ae8c-beb9dbce35ce","Type":"ContainerStarted","Data":"f98ff5241a039270a3b05822fea3c87f18859f74162897e6acb797aeca39838d"} Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.493855 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.493837363 podStartE2EDuration="3.493837363s" podCreationTimestamp="2025-10-05 07:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:13:32.491986594 +0000 UTC m=+1246.374613054" watchObservedRunningTime="2025-10-05 07:13:32.493837363 +0000 UTC m=+1246.376463823" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.507419 4935 scope.go:117] "RemoveContainer" containerID="9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4" Oct 05 07:13:32 crc kubenswrapper[4935]: W1005 07:13:32.515014 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5530001a_4323_49d7_adc3_eaa49ecb0963.slice/crio-9e3153f38ac0508a55c568091ebf10f12cd289c31f42ad0259735d8d92718788 WatchSource:0}: Error finding container 9e3153f38ac0508a55c568091ebf10f12cd289c31f42ad0259735d8d92718788: Status 404 returned error can't find the container with id 9e3153f38ac0508a55c568091ebf10f12cd289c31f42ad0259735d8d92718788 Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.524232 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.562212 4935 scope.go:117] "RemoveContainer" containerID="9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c" Oct 05 07:13:32 crc kubenswrapper[4935]: E1005 07:13:32.562785 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c\": container with ID starting with 9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c not found: ID does not exist" containerID="9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.562842 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c"} err="failed to get container status \"9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c\": rpc error: code = NotFound desc = could not find container \"9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c\": container with ID starting with 9f82c0fb54c6a6395db3a6acf16acbf5d834009bf768f09f1ae2e189e0a8189c not found: ID does not exist" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.562877 4935 scope.go:117] "RemoveContainer" containerID="9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4" Oct 05 07:13:32 crc kubenswrapper[4935]: E1005 07:13:32.563730 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4\": container with ID starting with 9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4 not found: ID does not exist" containerID="9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.563756 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4"} err="failed to get container status \"9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4\": rpc error: code = NotFound desc = could not find container \"9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4\": container with ID starting with 9782635d00898db399dd8b3608cd4562ecdc3e4f60e93b68f5d288d479708bd4 not found: ID does not exist" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582167 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-httpd-run\") pod \"b885bd46-8ccc-47d6-8622-c4175e2267e5\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582224 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-scripts\") pod \"b885bd46-8ccc-47d6-8622-c4175e2267e5\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582302 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knfb6\" (UniqueName: \"kubernetes.io/projected/b885bd46-8ccc-47d6-8622-c4175e2267e5-kube-api-access-knfb6\") pod \"b885bd46-8ccc-47d6-8622-c4175e2267e5\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582386 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-public-tls-certs\") pod \"b885bd46-8ccc-47d6-8622-c4175e2267e5\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582406 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b885bd46-8ccc-47d6-8622-c4175e2267e5\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582455 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-combined-ca-bundle\") pod \"b885bd46-8ccc-47d6-8622-c4175e2267e5\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582513 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-config-data\") pod \"b885bd46-8ccc-47d6-8622-c4175e2267e5\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582559 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-logs\") pod \"b885bd46-8ccc-47d6-8622-c4175e2267e5\" (UID: \"b885bd46-8ccc-47d6-8622-c4175e2267e5\") " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.582995 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b885bd46-8ccc-47d6-8622-c4175e2267e5" (UID: "b885bd46-8ccc-47d6-8622-c4175e2267e5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.583271 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.584873 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-logs" (OuterVolumeSpecName: "logs") pod "b885bd46-8ccc-47d6-8622-c4175e2267e5" (UID: "b885bd46-8ccc-47d6-8622-c4175e2267e5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.590106 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b885bd46-8ccc-47d6-8622-c4175e2267e5-kube-api-access-knfb6" (OuterVolumeSpecName: "kube-api-access-knfb6") pod "b885bd46-8ccc-47d6-8622-c4175e2267e5" (UID: "b885bd46-8ccc-47d6-8622-c4175e2267e5"). InnerVolumeSpecName "kube-api-access-knfb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.591990 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-scripts" (OuterVolumeSpecName: "scripts") pod "b885bd46-8ccc-47d6-8622-c4175e2267e5" (UID: "b885bd46-8ccc-47d6-8622-c4175e2267e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.593859 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "b885bd46-8ccc-47d6-8622-c4175e2267e5" (UID: "b885bd46-8ccc-47d6-8622-c4175e2267e5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.628078 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b885bd46-8ccc-47d6-8622-c4175e2267e5" (UID: "b885bd46-8ccc-47d6-8622-c4175e2267e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.652812 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-config-data" (OuterVolumeSpecName: "config-data") pod "b885bd46-8ccc-47d6-8622-c4175e2267e5" (UID: "b885bd46-8ccc-47d6-8622-c4175e2267e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.653652 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b885bd46-8ccc-47d6-8622-c4175e2267e5" (UID: "b885bd46-8ccc-47d6-8622-c4175e2267e5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.684914 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knfb6\" (UniqueName: \"kubernetes.io/projected/b885bd46-8ccc-47d6-8622-c4175e2267e5-kube-api-access-knfb6\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.684944 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.684974 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.684984 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.684994 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.685002 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b885bd46-8ccc-47d6-8622-c4175e2267e5-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.685012 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b885bd46-8ccc-47d6-8622-c4175e2267e5-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.705601 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.790099 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.793791 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be1be8df-b209-475f-b3cd-132e411f05f9" path="/var/lib/kubelet/pods/be1be8df-b209-475f-b3cd-132e411f05f9/volumes" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.820651 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.831020 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.848615 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:13:32 crc kubenswrapper[4935]: E1005 07:13:32.849199 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerName="glance-log" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.849220 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerName="glance-log" Oct 05 07:13:32 crc kubenswrapper[4935]: E1005 07:13:32.849246 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerName="glance-httpd" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.849256 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerName="glance-httpd" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.849486 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerName="glance-log" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.849523 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" containerName="glance-httpd" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.850813 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.853519 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.853763 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.871977 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.993706 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.994065 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-config-data\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.994102 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ssrz\" (UniqueName: \"kubernetes.io/projected/6c512e7c-9418-43b4-8235-4d62fa487556-kube-api-access-8ssrz\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.994145 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.994178 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.994197 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-scripts\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.994216 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-logs\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:32 crc kubenswrapper[4935]: I1005 07:13:32.994239 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.096429 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-config-data\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.096512 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ssrz\" (UniqueName: \"kubernetes.io/projected/6c512e7c-9418-43b4-8235-4d62fa487556-kube-api-access-8ssrz\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.096581 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.096634 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.096660 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-scripts\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.096688 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-logs\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.096718 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.096775 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.097125 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.097384 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.097523 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-logs\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.100865 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-scripts\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.102781 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.106501 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-config-data\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.108049 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.117037 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ssrz\" (UniqueName: \"kubernetes.io/projected/6c512e7c-9418-43b4-8235-4d62fa487556-kube-api-access-8ssrz\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.134479 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.187799 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.475683 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerStarted","Data":"3983b7c3f6ec782e452d427467eb3a5e402a2ab047cca535f2d430613b37cc6c"} Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.475988 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerStarted","Data":"9e3153f38ac0508a55c568091ebf10f12cd289c31f42ad0259735d8d92718788"} Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.736374 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.747853 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-70bb-account-create-2nf56"] Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.748980 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-70bb-account-create-2nf56" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.751369 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.759036 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-70bb-account-create-2nf56"] Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.814600 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6nbs\" (UniqueName: \"kubernetes.io/projected/319fa1f0-1aeb-4881-a883-4b2f7b61f833-kube-api-access-d6nbs\") pod \"nova-api-70bb-account-create-2nf56\" (UID: \"319fa1f0-1aeb-4881-a883-4b2f7b61f833\") " pod="openstack/nova-api-70bb-account-create-2nf56" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.917090 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6nbs\" (UniqueName: \"kubernetes.io/projected/319fa1f0-1aeb-4881-a883-4b2f7b61f833-kube-api-access-d6nbs\") pod \"nova-api-70bb-account-create-2nf56\" (UID: \"319fa1f0-1aeb-4881-a883-4b2f7b61f833\") " pod="openstack/nova-api-70bb-account-create-2nf56" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.929162 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9d1d-account-create-64df8"] Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.930287 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9d1d-account-create-64df8" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.932541 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.946712 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6nbs\" (UniqueName: \"kubernetes.io/projected/319fa1f0-1aeb-4881-a883-4b2f7b61f833-kube-api-access-d6nbs\") pod \"nova-api-70bb-account-create-2nf56\" (UID: \"319fa1f0-1aeb-4881-a883-4b2f7b61f833\") " pod="openstack/nova-api-70bb-account-create-2nf56" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.962476 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-70bb-account-create-2nf56" Oct 05 07:13:33 crc kubenswrapper[4935]: I1005 07:13:33.976146 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9d1d-account-create-64df8"] Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.021235 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5lzz\" (UniqueName: \"kubernetes.io/projected/f3fbc168-2981-4a9c-b4c4-0a6f1cec4556-kube-api-access-c5lzz\") pod \"nova-cell0-9d1d-account-create-64df8\" (UID: \"f3fbc168-2981-4a9c-b4c4-0a6f1cec4556\") " pod="openstack/nova-cell0-9d1d-account-create-64df8" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.126526 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5lzz\" (UniqueName: \"kubernetes.io/projected/f3fbc168-2981-4a9c-b4c4-0a6f1cec4556-kube-api-access-c5lzz\") pod \"nova-cell0-9d1d-account-create-64df8\" (UID: \"f3fbc168-2981-4a9c-b4c4-0a6f1cec4556\") " pod="openstack/nova-cell0-9d1d-account-create-64df8" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.135188 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1098-account-create-5zdbm"] Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.136365 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1098-account-create-5zdbm" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.140122 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.151211 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5lzz\" (UniqueName: \"kubernetes.io/projected/f3fbc168-2981-4a9c-b4c4-0a6f1cec4556-kube-api-access-c5lzz\") pod \"nova-cell0-9d1d-account-create-64df8\" (UID: \"f3fbc168-2981-4a9c-b4c4-0a6f1cec4556\") " pod="openstack/nova-cell0-9d1d-account-create-64df8" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.167963 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1098-account-create-5zdbm"] Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.228821 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2s4q\" (UniqueName: \"kubernetes.io/projected/94914bdc-6fed-4035-a624-93feacf3ba00-kube-api-access-z2s4q\") pod \"nova-cell1-1098-account-create-5zdbm\" (UID: \"94914bdc-6fed-4035-a624-93feacf3ba00\") " pod="openstack/nova-cell1-1098-account-create-5zdbm" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.318128 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9d1d-account-create-64df8" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.330910 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2s4q\" (UniqueName: \"kubernetes.io/projected/94914bdc-6fed-4035-a624-93feacf3ba00-kube-api-access-z2s4q\") pod \"nova-cell1-1098-account-create-5zdbm\" (UID: \"94914bdc-6fed-4035-a624-93feacf3ba00\") " pod="openstack/nova-cell1-1098-account-create-5zdbm" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.348930 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2s4q\" (UniqueName: \"kubernetes.io/projected/94914bdc-6fed-4035-a624-93feacf3ba00-kube-api-access-z2s4q\") pod \"nova-cell1-1098-account-create-5zdbm\" (UID: \"94914bdc-6fed-4035-a624-93feacf3ba00\") " pod="openstack/nova-cell1-1098-account-create-5zdbm" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.459395 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1098-account-create-5zdbm" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.511164 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-70bb-account-create-2nf56"] Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.511805 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c512e7c-9418-43b4-8235-4d62fa487556","Type":"ContainerStarted","Data":"83dd3f9b8ccfa6c15a5abedf1cc0dba702eb5c6c024d7e58ef7b5517611fcd66"} Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.511836 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c512e7c-9418-43b4-8235-4d62fa487556","Type":"ContainerStarted","Data":"a1ba0acdc2f2c344701c2fc076f77ad8459cad00319953c4929ff569e48699b4"} Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.523055 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerStarted","Data":"60d394c6136269bc33d60311928967097bad8f15fe68b127378448595e0c9e19"} Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.815482 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b885bd46-8ccc-47d6-8622-c4175e2267e5" path="/var/lib/kubelet/pods/b885bd46-8ccc-47d6-8622-c4175e2267e5/volumes" Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.926862 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9d1d-account-create-64df8"] Oct 05 07:13:34 crc kubenswrapper[4935]: I1005 07:13:34.961370 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1098-account-create-5zdbm"] Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.533866 4935 generic.go:334] "Generic (PLEG): container finished" podID="319fa1f0-1aeb-4881-a883-4b2f7b61f833" containerID="77c9c68a700241dead6230cd621003877e33deca266947dce8d521f5290a196a" exitCode=0 Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.533998 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-70bb-account-create-2nf56" event={"ID":"319fa1f0-1aeb-4881-a883-4b2f7b61f833","Type":"ContainerDied","Data":"77c9c68a700241dead6230cd621003877e33deca266947dce8d521f5290a196a"} Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.534202 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-70bb-account-create-2nf56" event={"ID":"319fa1f0-1aeb-4881-a883-4b2f7b61f833","Type":"ContainerStarted","Data":"58da7b37ea7638bcd03fd223f25710e768538122dd05ff220370d663db28d5d0"} Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.536434 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c512e7c-9418-43b4-8235-4d62fa487556","Type":"ContainerStarted","Data":"fc1c1a196142dfa9f9def44b621bd771b1f8fbde0b745b2687d81b122f138212"} Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.539148 4935 generic.go:334] "Generic (PLEG): container finished" podID="94914bdc-6fed-4035-a624-93feacf3ba00" containerID="ad1685789306deca0aceea8412449a0dfd2dbc9ebd426662244456257e54f051" exitCode=0 Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.539220 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1098-account-create-5zdbm" event={"ID":"94914bdc-6fed-4035-a624-93feacf3ba00","Type":"ContainerDied","Data":"ad1685789306deca0aceea8412449a0dfd2dbc9ebd426662244456257e54f051"} Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.539255 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1098-account-create-5zdbm" event={"ID":"94914bdc-6fed-4035-a624-93feacf3ba00","Type":"ContainerStarted","Data":"f1898a390c059042fa89913752a1ee56804d2c6e2dfc9cdd6d61d8c646ede40d"} Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.541177 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerStarted","Data":"5d6f6d0e2357a40b498816f830d4697c3a5a49a2193d341024dfe6fd920d8a8b"} Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.542546 4935 generic.go:334] "Generic (PLEG): container finished" podID="f3fbc168-2981-4a9c-b4c4-0a6f1cec4556" containerID="1226b9b0e256c9e620dc56bf185da59022a8cb075c026e4bc864757ee53d82ec" exitCode=0 Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.542571 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9d1d-account-create-64df8" event={"ID":"f3fbc168-2981-4a9c-b4c4-0a6f1cec4556","Type":"ContainerDied","Data":"1226b9b0e256c9e620dc56bf185da59022a8cb075c026e4bc864757ee53d82ec"} Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.542598 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9d1d-account-create-64df8" event={"ID":"f3fbc168-2981-4a9c-b4c4-0a6f1cec4556","Type":"ContainerStarted","Data":"c983401df219e08239809db78875a2be53dee456ae8b032850190f49e4ff4534"} Oct 05 07:13:35 crc kubenswrapper[4935]: I1005 07:13:35.593844 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.593821102 podStartE2EDuration="3.593821102s" podCreationTimestamp="2025-10-05 07:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:13:35.5899702 +0000 UTC m=+1249.472596660" watchObservedRunningTime="2025-10-05 07:13:35.593821102 +0000 UTC m=+1249.476447562" Oct 05 07:13:36 crc kubenswrapper[4935]: I1005 07:13:36.570855 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerStarted","Data":"325cfb275a728b4f50132ff8da8719b2d5a9ea238060405d3ff760c25c6c04a6"} Oct 05 07:13:36 crc kubenswrapper[4935]: I1005 07:13:36.600576 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.0107606 podStartE2EDuration="5.600541107s" podCreationTimestamp="2025-10-05 07:13:31 +0000 UTC" firstStartedPulling="2025-10-05 07:13:32.530035811 +0000 UTC m=+1246.412662271" lastFinishedPulling="2025-10-05 07:13:36.119816318 +0000 UTC m=+1250.002442778" observedRunningTime="2025-10-05 07:13:36.595605866 +0000 UTC m=+1250.478232346" watchObservedRunningTime="2025-10-05 07:13:36.600541107 +0000 UTC m=+1250.483167567" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.109774 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-70bb-account-create-2nf56" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.115264 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9d1d-account-create-64df8" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.120036 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1098-account-create-5zdbm" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.201596 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6nbs\" (UniqueName: \"kubernetes.io/projected/319fa1f0-1aeb-4881-a883-4b2f7b61f833-kube-api-access-d6nbs\") pod \"319fa1f0-1aeb-4881-a883-4b2f7b61f833\" (UID: \"319fa1f0-1aeb-4881-a883-4b2f7b61f833\") " Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.201737 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5lzz\" (UniqueName: \"kubernetes.io/projected/f3fbc168-2981-4a9c-b4c4-0a6f1cec4556-kube-api-access-c5lzz\") pod \"f3fbc168-2981-4a9c-b4c4-0a6f1cec4556\" (UID: \"f3fbc168-2981-4a9c-b4c4-0a6f1cec4556\") " Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.201790 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2s4q\" (UniqueName: \"kubernetes.io/projected/94914bdc-6fed-4035-a624-93feacf3ba00-kube-api-access-z2s4q\") pod \"94914bdc-6fed-4035-a624-93feacf3ba00\" (UID: \"94914bdc-6fed-4035-a624-93feacf3ba00\") " Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.207007 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3fbc168-2981-4a9c-b4c4-0a6f1cec4556-kube-api-access-c5lzz" (OuterVolumeSpecName: "kube-api-access-c5lzz") pod "f3fbc168-2981-4a9c-b4c4-0a6f1cec4556" (UID: "f3fbc168-2981-4a9c-b4c4-0a6f1cec4556"). InnerVolumeSpecName "kube-api-access-c5lzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.207781 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/319fa1f0-1aeb-4881-a883-4b2f7b61f833-kube-api-access-d6nbs" (OuterVolumeSpecName: "kube-api-access-d6nbs") pod "319fa1f0-1aeb-4881-a883-4b2f7b61f833" (UID: "319fa1f0-1aeb-4881-a883-4b2f7b61f833"). InnerVolumeSpecName "kube-api-access-d6nbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.219386 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94914bdc-6fed-4035-a624-93feacf3ba00-kube-api-access-z2s4q" (OuterVolumeSpecName: "kube-api-access-z2s4q") pod "94914bdc-6fed-4035-a624-93feacf3ba00" (UID: "94914bdc-6fed-4035-a624-93feacf3ba00"). InnerVolumeSpecName "kube-api-access-z2s4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.305175 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5lzz\" (UniqueName: \"kubernetes.io/projected/f3fbc168-2981-4a9c-b4c4-0a6f1cec4556-kube-api-access-c5lzz\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.305224 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2s4q\" (UniqueName: \"kubernetes.io/projected/94914bdc-6fed-4035-a624-93feacf3ba00-kube-api-access-z2s4q\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.305239 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6nbs\" (UniqueName: \"kubernetes.io/projected/319fa1f0-1aeb-4881-a883-4b2f7b61f833-kube-api-access-d6nbs\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.583131 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9d1d-account-create-64df8" event={"ID":"f3fbc168-2981-4a9c-b4c4-0a6f1cec4556","Type":"ContainerDied","Data":"c983401df219e08239809db78875a2be53dee456ae8b032850190f49e4ff4534"} Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.584011 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c983401df219e08239809db78875a2be53dee456ae8b032850190f49e4ff4534" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.584138 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9d1d-account-create-64df8" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.586886 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-70bb-account-create-2nf56" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.586926 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-70bb-account-create-2nf56" event={"ID":"319fa1f0-1aeb-4881-a883-4b2f7b61f833","Type":"ContainerDied","Data":"58da7b37ea7638bcd03fd223f25710e768538122dd05ff220370d663db28d5d0"} Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.586963 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58da7b37ea7638bcd03fd223f25710e768538122dd05ff220370d663db28d5d0" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.588296 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1098-account-create-5zdbm" event={"ID":"94914bdc-6fed-4035-a624-93feacf3ba00","Type":"ContainerDied","Data":"f1898a390c059042fa89913752a1ee56804d2c6e2dfc9cdd6d61d8c646ede40d"} Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.588341 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1898a390c059042fa89913752a1ee56804d2c6e2dfc9cdd6d61d8c646ede40d" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.588316 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1098-account-create-5zdbm" Oct 05 07:13:37 crc kubenswrapper[4935]: I1005 07:13:37.588562 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.133827 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxztf"] Oct 05 07:13:39 crc kubenswrapper[4935]: E1005 07:13:39.134707 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94914bdc-6fed-4035-a624-93feacf3ba00" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.134720 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="94914bdc-6fed-4035-a624-93feacf3ba00" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: E1005 07:13:39.134733 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fbc168-2981-4a9c-b4c4-0a6f1cec4556" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.134739 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fbc168-2981-4a9c-b4c4-0a6f1cec4556" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: E1005 07:13:39.134760 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319fa1f0-1aeb-4881-a883-4b2f7b61f833" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.134767 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="319fa1f0-1aeb-4881-a883-4b2f7b61f833" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.134950 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="319fa1f0-1aeb-4881-a883-4b2f7b61f833" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.134969 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="94914bdc-6fed-4035-a624-93feacf3ba00" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.134983 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fbc168-2981-4a9c-b4c4-0a6f1cec4556" containerName="mariadb-account-create" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.135585 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.140346 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pcjsx" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.140391 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.140534 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.147924 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxztf"] Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.238842 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-config-data\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.238886 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-scripts\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.238951 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.238987 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9v7c\" (UniqueName: \"kubernetes.io/projected/65623cb0-f66d-4bfd-b732-680ad4da5c1f-kube-api-access-s9v7c\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.340679 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-config-data\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.340721 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-scripts\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.340753 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.340779 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9v7c\" (UniqueName: \"kubernetes.io/projected/65623cb0-f66d-4bfd-b732-680ad4da5c1f-kube-api-access-s9v7c\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.345544 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-scripts\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.349594 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-config-data\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.358020 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.366554 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9v7c\" (UniqueName: \"kubernetes.io/projected/65623cb0-f66d-4bfd-b732-680ad4da5c1f-kube-api-access-s9v7c\") pod \"nova-cell0-conductor-db-sync-wxztf\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.484198 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.746666 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxztf"] Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.776903 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.776942 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.813357 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:39 crc kubenswrapper[4935]: I1005 07:13:39.821775 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:40 crc kubenswrapper[4935]: I1005 07:13:40.039402 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 07:13:40 crc kubenswrapper[4935]: I1005 07:13:40.624836 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wxztf" event={"ID":"65623cb0-f66d-4bfd-b732-680ad4da5c1f","Type":"ContainerStarted","Data":"39f5e9053aaffab69e3bedc0810fa791994ef602b92f6177ea97f522dbaea908"} Oct 05 07:13:40 crc kubenswrapper[4935]: I1005 07:13:40.624911 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:40 crc kubenswrapper[4935]: I1005 07:13:40.624925 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:42 crc kubenswrapper[4935]: I1005 07:13:42.643175 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 07:13:42 crc kubenswrapper[4935]: I1005 07:13:42.654235 4935 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 07:13:42 crc kubenswrapper[4935]: I1005 07:13:42.859555 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:42 crc kubenswrapper[4935]: I1005 07:13:42.860412 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.146157 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.146454 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="ceilometer-central-agent" containerID="cri-o://3983b7c3f6ec782e452d427467eb3a5e402a2ab047cca535f2d430613b37cc6c" gracePeriod=30 Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.147116 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="proxy-httpd" containerID="cri-o://325cfb275a728b4f50132ff8da8719b2d5a9ea238060405d3ff760c25c6c04a6" gracePeriod=30 Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.147184 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="sg-core" containerID="cri-o://5d6f6d0e2357a40b498816f830d4697c3a5a49a2193d341024dfe6fd920d8a8b" gracePeriod=30 Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.147233 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="ceilometer-notification-agent" containerID="cri-o://60d394c6136269bc33d60311928967097bad8f15fe68b127378448595e0c9e19" gracePeriod=30 Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.189273 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.189567 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.229708 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.233709 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.653011 4935 generic.go:334] "Generic (PLEG): container finished" podID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerID="325cfb275a728b4f50132ff8da8719b2d5a9ea238060405d3ff760c25c6c04a6" exitCode=0 Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.653296 4935 generic.go:334] "Generic (PLEG): container finished" podID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerID="5d6f6d0e2357a40b498816f830d4697c3a5a49a2193d341024dfe6fd920d8a8b" exitCode=2 Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.653081 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerDied","Data":"325cfb275a728b4f50132ff8da8719b2d5a9ea238060405d3ff760c25c6c04a6"} Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.653337 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerDied","Data":"5d6f6d0e2357a40b498816f830d4697c3a5a49a2193d341024dfe6fd920d8a8b"} Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.653350 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerDied","Data":"3983b7c3f6ec782e452d427467eb3a5e402a2ab047cca535f2d430613b37cc6c"} Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.653306 4935 generic.go:334] "Generic (PLEG): container finished" podID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerID="3983b7c3f6ec782e452d427467eb3a5e402a2ab047cca535f2d430613b37cc6c" exitCode=0 Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.654008 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 07:13:43 crc kubenswrapper[4935]: I1005 07:13:43.654059 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 07:13:44 crc kubenswrapper[4935]: I1005 07:13:44.289751 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:13:44 crc kubenswrapper[4935]: I1005 07:13:44.289807 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:13:44 crc kubenswrapper[4935]: I1005 07:13:44.666250 4935 generic.go:334] "Generic (PLEG): container finished" podID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerID="60d394c6136269bc33d60311928967097bad8f15fe68b127378448595e0c9e19" exitCode=0 Oct 05 07:13:44 crc kubenswrapper[4935]: I1005 07:13:44.667457 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerDied","Data":"60d394c6136269bc33d60311928967097bad8f15fe68b127378448595e0c9e19"} Oct 05 07:13:45 crc kubenswrapper[4935]: I1005 07:13:45.480678 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 07:13:45 crc kubenswrapper[4935]: I1005 07:13:45.512270 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 07:13:48 crc kubenswrapper[4935]: I1005 07:13:48.946813 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.023566 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-sg-core-conf-yaml\") pod \"5530001a-4323-49d7-adc3-eaa49ecb0963\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.023643 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-log-httpd\") pod \"5530001a-4323-49d7-adc3-eaa49ecb0963\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.023671 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-scripts\") pod \"5530001a-4323-49d7-adc3-eaa49ecb0963\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.023770 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-config-data\") pod \"5530001a-4323-49d7-adc3-eaa49ecb0963\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.023883 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-run-httpd\") pod \"5530001a-4323-49d7-adc3-eaa49ecb0963\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.023956 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b5mp\" (UniqueName: \"kubernetes.io/projected/5530001a-4323-49d7-adc3-eaa49ecb0963-kube-api-access-7b5mp\") pod \"5530001a-4323-49d7-adc3-eaa49ecb0963\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.023980 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-combined-ca-bundle\") pod \"5530001a-4323-49d7-adc3-eaa49ecb0963\" (UID: \"5530001a-4323-49d7-adc3-eaa49ecb0963\") " Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.024346 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5530001a-4323-49d7-adc3-eaa49ecb0963" (UID: "5530001a-4323-49d7-adc3-eaa49ecb0963"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.024800 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5530001a-4323-49d7-adc3-eaa49ecb0963" (UID: "5530001a-4323-49d7-adc3-eaa49ecb0963"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.028180 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5530001a-4323-49d7-adc3-eaa49ecb0963-kube-api-access-7b5mp" (OuterVolumeSpecName: "kube-api-access-7b5mp") pod "5530001a-4323-49d7-adc3-eaa49ecb0963" (UID: "5530001a-4323-49d7-adc3-eaa49ecb0963"). InnerVolumeSpecName "kube-api-access-7b5mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.028279 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-scripts" (OuterVolumeSpecName: "scripts") pod "5530001a-4323-49d7-adc3-eaa49ecb0963" (UID: "5530001a-4323-49d7-adc3-eaa49ecb0963"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.048023 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5530001a-4323-49d7-adc3-eaa49ecb0963" (UID: "5530001a-4323-49d7-adc3-eaa49ecb0963"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.119719 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5530001a-4323-49d7-adc3-eaa49ecb0963" (UID: "5530001a-4323-49d7-adc3-eaa49ecb0963"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.122176 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-config-data" (OuterVolumeSpecName: "config-data") pod "5530001a-4323-49d7-adc3-eaa49ecb0963" (UID: "5530001a-4323-49d7-adc3-eaa49ecb0963"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.126682 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.126711 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.126719 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.126727 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.126734 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5530001a-4323-49d7-adc3-eaa49ecb0963-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.126744 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b5mp\" (UniqueName: \"kubernetes.io/projected/5530001a-4323-49d7-adc3-eaa49ecb0963-kube-api-access-7b5mp\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.126753 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5530001a-4323-49d7-adc3-eaa49ecb0963-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.726863 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wxztf" event={"ID":"65623cb0-f66d-4bfd-b732-680ad4da5c1f","Type":"ContainerStarted","Data":"6785a03b97344eb642561ee5d12730a7f93de79453bf17b263bd20ae49bbd3b3"} Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.732255 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5530001a-4323-49d7-adc3-eaa49ecb0963","Type":"ContainerDied","Data":"9e3153f38ac0508a55c568091ebf10f12cd289c31f42ad0259735d8d92718788"} Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.732288 4935 scope.go:117] "RemoveContainer" containerID="325cfb275a728b4f50132ff8da8719b2d5a9ea238060405d3ff760c25c6c04a6" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.732403 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.749286 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-wxztf" podStartSLOduration=1.788805187 podStartE2EDuration="10.74926482s" podCreationTimestamp="2025-10-05 07:13:39 +0000 UTC" firstStartedPulling="2025-10-05 07:13:39.757257915 +0000 UTC m=+1253.639884375" lastFinishedPulling="2025-10-05 07:13:48.717717548 +0000 UTC m=+1262.600344008" observedRunningTime="2025-10-05 07:13:49.742716967 +0000 UTC m=+1263.625343427" watchObservedRunningTime="2025-10-05 07:13:49.74926482 +0000 UTC m=+1263.631891280" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.761950 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.768710 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.769936 4935 scope.go:117] "RemoveContainer" containerID="5d6f6d0e2357a40b498816f830d4697c3a5a49a2193d341024dfe6fd920d8a8b" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792131 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:49 crc kubenswrapper[4935]: E1005 07:13:49.792467 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="sg-core" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792483 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="sg-core" Oct 05 07:13:49 crc kubenswrapper[4935]: E1005 07:13:49.792510 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="ceilometer-central-agent" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792516 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="ceilometer-central-agent" Oct 05 07:13:49 crc kubenswrapper[4935]: E1005 07:13:49.792538 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="proxy-httpd" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792543 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="proxy-httpd" Oct 05 07:13:49 crc kubenswrapper[4935]: E1005 07:13:49.792558 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="ceilometer-notification-agent" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792566 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="ceilometer-notification-agent" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792721 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="proxy-httpd" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792730 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="ceilometer-notification-agent" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792747 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="sg-core" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.792760 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" containerName="ceilometer-central-agent" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.794259 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.808460 4935 scope.go:117] "RemoveContainer" containerID="60d394c6136269bc33d60311928967097bad8f15fe68b127378448595e0c9e19" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.808619 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.808802 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.816713 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.847839 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.847923 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-log-httpd\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.847995 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-config-data\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.848038 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-scripts\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.848078 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlh2n\" (UniqueName: \"kubernetes.io/projected/b95f4dea-2474-4502-a492-f9ceea8df986-kube-api-access-mlh2n\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.848096 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.848124 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-run-httpd\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.871578 4935 scope.go:117] "RemoveContainer" containerID="3983b7c3f6ec782e452d427467eb3a5e402a2ab047cca535f2d430613b37cc6c" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.949829 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.949914 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-log-httpd\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.949961 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-config-data\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.950012 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-scripts\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.950069 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlh2n\" (UniqueName: \"kubernetes.io/projected/b95f4dea-2474-4502-a492-f9ceea8df986-kube-api-access-mlh2n\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.950093 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.950129 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-run-httpd\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.950464 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-log-httpd\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.950518 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-run-httpd\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.957122 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.957130 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-scripts\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.959821 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.965496 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-config-data\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:49 crc kubenswrapper[4935]: I1005 07:13:49.974148 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlh2n\" (UniqueName: \"kubernetes.io/projected/b95f4dea-2474-4502-a492-f9ceea8df986-kube-api-access-mlh2n\") pod \"ceilometer-0\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " pod="openstack/ceilometer-0" Oct 05 07:13:50 crc kubenswrapper[4935]: I1005 07:13:50.160370 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:13:50 crc kubenswrapper[4935]: I1005 07:13:50.644540 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:50 crc kubenswrapper[4935]: I1005 07:13:50.740459 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerStarted","Data":"23437712c7502b96f0846ddc83e84815738d6f60df6d8a45649da8bdf2cefb66"} Oct 05 07:13:50 crc kubenswrapper[4935]: I1005 07:13:50.788549 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5530001a-4323-49d7-adc3-eaa49ecb0963" path="/var/lib/kubelet/pods/5530001a-4323-49d7-adc3-eaa49ecb0963/volumes" Oct 05 07:13:51 crc kubenswrapper[4935]: I1005 07:13:51.750432 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerStarted","Data":"98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514"} Oct 05 07:13:52 crc kubenswrapper[4935]: I1005 07:13:52.256454 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:13:52 crc kubenswrapper[4935]: I1005 07:13:52.761109 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerStarted","Data":"343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198"} Oct 05 07:13:53 crc kubenswrapper[4935]: I1005 07:13:53.781187 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerStarted","Data":"9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d"} Oct 05 07:13:54 crc kubenswrapper[4935]: I1005 07:13:54.809169 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerStarted","Data":"0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126"} Oct 05 07:13:54 crc kubenswrapper[4935]: I1005 07:13:54.809367 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="ceilometer-central-agent" containerID="cri-o://98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514" gracePeriod=30 Oct 05 07:13:54 crc kubenswrapper[4935]: I1005 07:13:54.809609 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:13:54 crc kubenswrapper[4935]: I1005 07:13:54.809677 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="proxy-httpd" containerID="cri-o://0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126" gracePeriod=30 Oct 05 07:13:54 crc kubenswrapper[4935]: I1005 07:13:54.809735 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="ceilometer-notification-agent" containerID="cri-o://343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198" gracePeriod=30 Oct 05 07:13:54 crc kubenswrapper[4935]: I1005 07:13:54.809738 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="sg-core" containerID="cri-o://9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d" gracePeriod=30 Oct 05 07:13:54 crc kubenswrapper[4935]: I1005 07:13:54.831588 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.16288108 podStartE2EDuration="5.831570925s" podCreationTimestamp="2025-10-05 07:13:49 +0000 UTC" firstStartedPulling="2025-10-05 07:13:50.648193403 +0000 UTC m=+1264.530819863" lastFinishedPulling="2025-10-05 07:13:54.316883238 +0000 UTC m=+1268.199509708" observedRunningTime="2025-10-05 07:13:54.829393687 +0000 UTC m=+1268.712020157" watchObservedRunningTime="2025-10-05 07:13:54.831570925 +0000 UTC m=+1268.714197395" Oct 05 07:13:55 crc kubenswrapper[4935]: I1005 07:13:55.823393 4935 generic.go:334] "Generic (PLEG): container finished" podID="b95f4dea-2474-4502-a492-f9ceea8df986" containerID="0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126" exitCode=0 Oct 05 07:13:55 crc kubenswrapper[4935]: I1005 07:13:55.823759 4935 generic.go:334] "Generic (PLEG): container finished" podID="b95f4dea-2474-4502-a492-f9ceea8df986" containerID="9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d" exitCode=2 Oct 05 07:13:55 crc kubenswrapper[4935]: I1005 07:13:55.823768 4935 generic.go:334] "Generic (PLEG): container finished" podID="b95f4dea-2474-4502-a492-f9ceea8df986" containerID="343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198" exitCode=0 Oct 05 07:13:55 crc kubenswrapper[4935]: I1005 07:13:55.823461 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerDied","Data":"0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126"} Oct 05 07:13:55 crc kubenswrapper[4935]: I1005 07:13:55.823805 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerDied","Data":"9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d"} Oct 05 07:13:55 crc kubenswrapper[4935]: I1005 07:13:55.823821 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerDied","Data":"343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198"} Oct 05 07:13:58 crc kubenswrapper[4935]: I1005 07:13:58.863517 4935 generic.go:334] "Generic (PLEG): container finished" podID="65623cb0-f66d-4bfd-b732-680ad4da5c1f" containerID="6785a03b97344eb642561ee5d12730a7f93de79453bf17b263bd20ae49bbd3b3" exitCode=0 Oct 05 07:13:58 crc kubenswrapper[4935]: I1005 07:13:58.863609 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wxztf" event={"ID":"65623cb0-f66d-4bfd-b732-680ad4da5c1f","Type":"ContainerDied","Data":"6785a03b97344eb642561ee5d12730a7f93de79453bf17b263bd20ae49bbd3b3"} Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.297256 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.347769 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9v7c\" (UniqueName: \"kubernetes.io/projected/65623cb0-f66d-4bfd-b732-680ad4da5c1f-kube-api-access-s9v7c\") pod \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.347886 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-combined-ca-bundle\") pod \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.348835 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-config-data\") pod \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.348901 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-scripts\") pod \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\" (UID: \"65623cb0-f66d-4bfd-b732-680ad4da5c1f\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.353392 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-scripts" (OuterVolumeSpecName: "scripts") pod "65623cb0-f66d-4bfd-b732-680ad4da5c1f" (UID: "65623cb0-f66d-4bfd-b732-680ad4da5c1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.366395 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65623cb0-f66d-4bfd-b732-680ad4da5c1f-kube-api-access-s9v7c" (OuterVolumeSpecName: "kube-api-access-s9v7c") pod "65623cb0-f66d-4bfd-b732-680ad4da5c1f" (UID: "65623cb0-f66d-4bfd-b732-680ad4da5c1f"). InnerVolumeSpecName "kube-api-access-s9v7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.382523 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65623cb0-f66d-4bfd-b732-680ad4da5c1f" (UID: "65623cb0-f66d-4bfd-b732-680ad4da5c1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.395349 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-config-data" (OuterVolumeSpecName: "config-data") pod "65623cb0-f66d-4bfd-b732-680ad4da5c1f" (UID: "65623cb0-f66d-4bfd-b732-680ad4da5c1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.451096 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9v7c\" (UniqueName: \"kubernetes.io/projected/65623cb0-f66d-4bfd-b732-680ad4da5c1f-kube-api-access-s9v7c\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.451128 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.451139 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.451147 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65623cb0-f66d-4bfd-b732-680ad4da5c1f-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.626178 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.792725 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlh2n\" (UniqueName: \"kubernetes.io/projected/b95f4dea-2474-4502-a492-f9ceea8df986-kube-api-access-mlh2n\") pod \"b95f4dea-2474-4502-a492-f9ceea8df986\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.792815 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-sg-core-conf-yaml\") pod \"b95f4dea-2474-4502-a492-f9ceea8df986\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.792887 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-run-httpd\") pod \"b95f4dea-2474-4502-a492-f9ceea8df986\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.792935 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-log-httpd\") pod \"b95f4dea-2474-4502-a492-f9ceea8df986\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.792982 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-config-data\") pod \"b95f4dea-2474-4502-a492-f9ceea8df986\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.793037 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-combined-ca-bundle\") pod \"b95f4dea-2474-4502-a492-f9ceea8df986\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.793079 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-scripts\") pod \"b95f4dea-2474-4502-a492-f9ceea8df986\" (UID: \"b95f4dea-2474-4502-a492-f9ceea8df986\") " Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.793291 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b95f4dea-2474-4502-a492-f9ceea8df986" (UID: "b95f4dea-2474-4502-a492-f9ceea8df986"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.793418 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b95f4dea-2474-4502-a492-f9ceea8df986" (UID: "b95f4dea-2474-4502-a492-f9ceea8df986"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.793812 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.793838 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b95f4dea-2474-4502-a492-f9ceea8df986-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.796397 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b95f4dea-2474-4502-a492-f9ceea8df986-kube-api-access-mlh2n" (OuterVolumeSpecName: "kube-api-access-mlh2n") pod "b95f4dea-2474-4502-a492-f9ceea8df986" (UID: "b95f4dea-2474-4502-a492-f9ceea8df986"). InnerVolumeSpecName "kube-api-access-mlh2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.797926 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-scripts" (OuterVolumeSpecName: "scripts") pod "b95f4dea-2474-4502-a492-f9ceea8df986" (UID: "b95f4dea-2474-4502-a492-f9ceea8df986"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.818376 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b95f4dea-2474-4502-a492-f9ceea8df986" (UID: "b95f4dea-2474-4502-a492-f9ceea8df986"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.869857 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b95f4dea-2474-4502-a492-f9ceea8df986" (UID: "b95f4dea-2474-4502-a492-f9ceea8df986"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.884139 4935 generic.go:334] "Generic (PLEG): container finished" podID="b95f4dea-2474-4502-a492-f9ceea8df986" containerID="98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514" exitCode=0 Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.884235 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerDied","Data":"98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514"} Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.884255 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.884284 4935 scope.go:117] "RemoveContainer" containerID="0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.884271 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b95f4dea-2474-4502-a492-f9ceea8df986","Type":"ContainerDied","Data":"23437712c7502b96f0846ddc83e84815738d6f60df6d8a45649da8bdf2cefb66"} Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.888329 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wxztf" event={"ID":"65623cb0-f66d-4bfd-b732-680ad4da5c1f","Type":"ContainerDied","Data":"39f5e9053aaffab69e3bedc0810fa791994ef602b92f6177ea97f522dbaea908"} Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.888381 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39f5e9053aaffab69e3bedc0810fa791994ef602b92f6177ea97f522dbaea908" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.888460 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wxztf" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.895528 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.895565 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.895584 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlh2n\" (UniqueName: \"kubernetes.io/projected/b95f4dea-2474-4502-a492-f9ceea8df986-kube-api-access-mlh2n\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.895601 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.911355 4935 scope.go:117] "RemoveContainer" containerID="9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.914371 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-config-data" (OuterVolumeSpecName: "config-data") pod "b95f4dea-2474-4502-a492-f9ceea8df986" (UID: "b95f4dea-2474-4502-a492-f9ceea8df986"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.949132 4935 scope.go:117] "RemoveContainer" containerID="343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.981294 4935 scope.go:117] "RemoveContainer" containerID="98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514" Oct 05 07:14:00 crc kubenswrapper[4935]: I1005 07:14:00.998209 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b95f4dea-2474-4502-a492-f9ceea8df986-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.023650 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.024180 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="ceilometer-central-agent" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.024279 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="ceilometer-central-agent" Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.024373 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="sg-core" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.024430 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="sg-core" Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.024491 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="ceilometer-notification-agent" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.024548 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="ceilometer-notification-agent" Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.024606 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65623cb0-f66d-4bfd-b732-680ad4da5c1f" containerName="nova-cell0-conductor-db-sync" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.024661 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="65623cb0-f66d-4bfd-b732-680ad4da5c1f" containerName="nova-cell0-conductor-db-sync" Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.024722 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="proxy-httpd" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.024779 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="proxy-httpd" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.025000 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="ceilometer-notification-agent" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.025075 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="proxy-httpd" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.025142 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="ceilometer-central-agent" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.025203 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="65623cb0-f66d-4bfd-b732-680ad4da5c1f" containerName="nova-cell0-conductor-db-sync" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.025263 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" containerName="sg-core" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.025805 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.025942 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.030588 4935 scope.go:117] "RemoveContainer" containerID="0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.030801 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.030932 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pcjsx" Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.032205 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126\": container with ID starting with 0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126 not found: ID does not exist" containerID="0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.032234 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126"} err="failed to get container status \"0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126\": rpc error: code = NotFound desc = could not find container \"0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126\": container with ID starting with 0ece466eee6d384dd28e19554d53bfb9a6db2810958b55aa66c50972ac32c126 not found: ID does not exist" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.032254 4935 scope.go:117] "RemoveContainer" containerID="9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d" Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.033597 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d\": container with ID starting with 9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d not found: ID does not exist" containerID="9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.033622 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d"} err="failed to get container status \"9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d\": rpc error: code = NotFound desc = could not find container \"9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d\": container with ID starting with 9e75840e8f1ccc046092c957919d91c64a07913fe27408024889f1c45faea99d not found: ID does not exist" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.033635 4935 scope.go:117] "RemoveContainer" containerID="343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198" Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.035677 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198\": container with ID starting with 343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198 not found: ID does not exist" containerID="343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.035738 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198"} err="failed to get container status \"343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198\": rpc error: code = NotFound desc = could not find container \"343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198\": container with ID starting with 343b9073c2bd6b359dc7063f5b21cb2b627b124c21d584fa1532d366e6a86198 not found: ID does not exist" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.035759 4935 scope.go:117] "RemoveContainer" containerID="98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514" Oct 05 07:14:01 crc kubenswrapper[4935]: E1005 07:14:01.036966 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514\": container with ID starting with 98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514 not found: ID does not exist" containerID="98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.037072 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514"} err="failed to get container status \"98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514\": rpc error: code = NotFound desc = could not find container \"98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514\": container with ID starting with 98f71bb552c2de06ffd414fd4e46c4e90f3ffe5727607cab2054c91d47451514 not found: ID does not exist" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.099464 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.099511 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89kfs\" (UniqueName: \"kubernetes.io/projected/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-kube-api-access-89kfs\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.099604 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.200708 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.200756 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89kfs\" (UniqueName: \"kubernetes.io/projected/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-kube-api-access-89kfs\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.200831 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.205225 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.206228 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.228997 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.233403 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89kfs\" (UniqueName: \"kubernetes.io/projected/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-kube-api-access-89kfs\") pod \"nova-cell0-conductor-0\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.239113 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.260075 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.262158 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.265787 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.265984 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.278597 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.345330 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.405786 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.405845 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-config-data\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.405933 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwczs\" (UniqueName: \"kubernetes.io/projected/86d0c526-9ff7-4418-a237-92ab6a75b224-kube-api-access-hwczs\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.405956 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-run-httpd\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.406041 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.406072 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-log-httpd\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.406123 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-scripts\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.510198 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-scripts\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.510637 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.510670 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-config-data\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.510766 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwczs\" (UniqueName: \"kubernetes.io/projected/86d0c526-9ff7-4418-a237-92ab6a75b224-kube-api-access-hwczs\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.510819 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-run-httpd\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.511019 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.511182 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-log-httpd\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.513187 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-run-httpd\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.513202 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-log-httpd\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.516704 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.517152 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.518385 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-config-data\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.527339 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-scripts\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.529674 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwczs\" (UniqueName: \"kubernetes.io/projected/86d0c526-9ff7-4418-a237-92ab6a75b224-kube-api-access-hwczs\") pod \"ceilometer-0\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.666089 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.792395 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:14:01 crc kubenswrapper[4935]: W1005 07:14:01.798565 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f2e339d_5a3f_4ab4_91a3_3fc73a96f285.slice/crio-080679fceb913252b8ca0d46672b01abb1ff18b9b01f1e11fa93f71b02709f61 WatchSource:0}: Error finding container 080679fceb913252b8ca0d46672b01abb1ff18b9b01f1e11fa93f71b02709f61: Status 404 returned error can't find the container with id 080679fceb913252b8ca0d46672b01abb1ff18b9b01f1e11fa93f71b02709f61 Oct 05 07:14:01 crc kubenswrapper[4935]: I1005 07:14:01.899230 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285","Type":"ContainerStarted","Data":"080679fceb913252b8ca0d46672b01abb1ff18b9b01f1e11fa93f71b02709f61"} Oct 05 07:14:02 crc kubenswrapper[4935]: I1005 07:14:02.110617 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:02 crc kubenswrapper[4935]: W1005 07:14:02.113433 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86d0c526_9ff7_4418_a237_92ab6a75b224.slice/crio-7c044fe27b930c731a2e7a5d53f6c993308c009d5090fa1814f9075cfd63c1f5 WatchSource:0}: Error finding container 7c044fe27b930c731a2e7a5d53f6c993308c009d5090fa1814f9075cfd63c1f5: Status 404 returned error can't find the container with id 7c044fe27b930c731a2e7a5d53f6c993308c009d5090fa1814f9075cfd63c1f5 Oct 05 07:14:02 crc kubenswrapper[4935]: I1005 07:14:02.791938 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b95f4dea-2474-4502-a492-f9ceea8df986" path="/var/lib/kubelet/pods/b95f4dea-2474-4502-a492-f9ceea8df986/volumes" Oct 05 07:14:02 crc kubenswrapper[4935]: I1005 07:14:02.909830 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerStarted","Data":"692ed475fb3453a3e6952fec05183b2dc6f37080fcd7dbbaf53b0d0a2e32e4ac"} Oct 05 07:14:02 crc kubenswrapper[4935]: I1005 07:14:02.910233 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerStarted","Data":"7c044fe27b930c731a2e7a5d53f6c993308c009d5090fa1814f9075cfd63c1f5"} Oct 05 07:14:02 crc kubenswrapper[4935]: I1005 07:14:02.911395 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285","Type":"ContainerStarted","Data":"24a61e5d809f2202e8aff2ffc919b4f809f3ea3e05132213aaafa01f0e1fc0d4"} Oct 05 07:14:02 crc kubenswrapper[4935]: I1005 07:14:02.911504 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:02 crc kubenswrapper[4935]: I1005 07:14:02.932967 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.932950688 podStartE2EDuration="2.932950688s" podCreationTimestamp="2025-10-05 07:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:02.929151508 +0000 UTC m=+1276.811777978" watchObservedRunningTime="2025-10-05 07:14:02.932950688 +0000 UTC m=+1276.815577148" Oct 05 07:14:03 crc kubenswrapper[4935]: I1005 07:14:03.931791 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerStarted","Data":"ed56f6247798e566f4409b29c17aa572f76c05d02e825db11fb0cc73b7664132"} Oct 05 07:14:04 crc kubenswrapper[4935]: I1005 07:14:04.951276 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerStarted","Data":"345c0040604dcbfb3be52da65117f237ffd6be39c38f28fb647f162dc6931d3f"} Oct 05 07:14:05 crc kubenswrapper[4935]: I1005 07:14:05.961353 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerStarted","Data":"696aa393f941516e4cfa1b451dd2f3e4440d4dee11149b8365d4707e731a8c95"} Oct 05 07:14:05 crc kubenswrapper[4935]: I1005 07:14:05.961946 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:14:05 crc kubenswrapper[4935]: I1005 07:14:05.984669 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.037240798 podStartE2EDuration="4.984646328s" podCreationTimestamp="2025-10-05 07:14:01 +0000 UTC" firstStartedPulling="2025-10-05 07:14:02.115741087 +0000 UTC m=+1275.998367547" lastFinishedPulling="2025-10-05 07:14:05.063146587 +0000 UTC m=+1278.945773077" observedRunningTime="2025-10-05 07:14:05.984387331 +0000 UTC m=+1279.867013811" watchObservedRunningTime="2025-10-05 07:14:05.984646328 +0000 UTC m=+1279.867272788" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.369509 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.821652 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-4w7kp"] Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.823082 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.828592 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.828603 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.833281 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4w7kp"] Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.923995 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.924126 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj9q6\" (UniqueName: \"kubernetes.io/projected/c2def690-05d7-4196-b383-58669b274802-kube-api-access-wj9q6\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.924168 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-scripts\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.924246 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-config-data\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.955513 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.956997 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.958918 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 07:14:06 crc kubenswrapper[4935]: I1005 07:14:06.974802 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.016699 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.038874 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.041714 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.052253 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.078873 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.079183 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj9q6\" (UniqueName: \"kubernetes.io/projected/c2def690-05d7-4196-b383-58669b274802-kube-api-access-wj9q6\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.079310 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-scripts\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.079505 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-config-data\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.089865 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-config-data\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.096191 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-scripts\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.108883 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.156197 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj9q6\" (UniqueName: \"kubernetes.io/projected/c2def690-05d7-4196-b383-58669b274802-kube-api-access-wj9q6\") pod \"nova-cell0-cell-mapping-4w7kp\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.160962 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-5qg9s"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.162554 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.181250 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.182688 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.187730 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.187968 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-swift-storage-0\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188061 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-config-data\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188144 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n96vp\" (UniqueName: \"kubernetes.io/projected/a02edeb1-77e8-4884-a922-3dc7d608b587-kube-api-access-n96vp\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188261 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chxcv\" (UniqueName: \"kubernetes.io/projected/b51b1cb3-ea91-4535-833b-03138c9c9eb4-kube-api-access-chxcv\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188339 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-svc\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188422 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv5vp\" (UniqueName: \"kubernetes.io/projected/59688ed9-ff1f-4c93-961c-247789e17af0-kube-api-access-rv5vp\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188508 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188606 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188704 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-logs\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188776 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-config\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.188936 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.189040 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59688ed9-ff1f-4c93-961c-247789e17af0-logs\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.189128 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64tgc\" (UniqueName: \"kubernetes.io/projected/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-kube-api-access-64tgc\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.189304 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-config-data\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.189394 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.189471 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-config-data\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.191461 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.202839 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.206963 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.231953 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-5qg9s"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.270668 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.278372 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.288432 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291166 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291295 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-config-data\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291404 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291489 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-swift-storage-0\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291569 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-config-data\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291642 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n96vp\" (UniqueName: \"kubernetes.io/projected/a02edeb1-77e8-4884-a922-3dc7d608b587-kube-api-access-n96vp\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291720 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chxcv\" (UniqueName: \"kubernetes.io/projected/b51b1cb3-ea91-4535-833b-03138c9c9eb4-kube-api-access-chxcv\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291800 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-svc\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291873 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv5vp\" (UniqueName: \"kubernetes.io/projected/59688ed9-ff1f-4c93-961c-247789e17af0-kube-api-access-rv5vp\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.291982 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.292067 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.292147 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-logs\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.292217 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-config\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.292316 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.292401 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59688ed9-ff1f-4c93-961c-247789e17af0-logs\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.295591 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64tgc\" (UniqueName: \"kubernetes.io/projected/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-kube-api-access-64tgc\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.295749 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-config-data\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.296269 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-logs\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.296577 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-swift-storage-0\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.296855 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-svc\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.297355 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.306100 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59688ed9-ff1f-4c93-961c-247789e17af0-logs\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.307732 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.292263 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.308361 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-config\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.310991 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.315446 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-config-data\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.316498 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.318015 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chxcv\" (UniqueName: \"kubernetes.io/projected/b51b1cb3-ea91-4535-833b-03138c9c9eb4-kube-api-access-chxcv\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.328371 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-config-data\") pod \"nova-scheduler-0\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.328611 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv5vp\" (UniqueName: \"kubernetes.io/projected/59688ed9-ff1f-4c93-961c-247789e17af0-kube-api-access-rv5vp\") pod \"nova-api-0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.329370 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-config-data\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.337256 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n96vp\" (UniqueName: \"kubernetes.io/projected/a02edeb1-77e8-4884-a922-3dc7d608b587-kube-api-access-n96vp\") pod \"dnsmasq-dns-7d6d46f6cf-5qg9s\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.338636 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64tgc\" (UniqueName: \"kubernetes.io/projected/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-kube-api-access-64tgc\") pod \"nova-metadata-0\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.360960 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.409821 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.410451 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbk5b\" (UniqueName: \"kubernetes.io/projected/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-kube-api-access-pbk5b\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.410642 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.415478 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.427079 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.503328 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.512657 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.512754 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbk5b\" (UniqueName: \"kubernetes.io/projected/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-kube-api-access-pbk5b\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.512818 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.520752 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.535845 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.538511 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbk5b\" (UniqueName: \"kubernetes.io/projected/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-kube-api-access-pbk5b\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.573631 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.747323 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.834339 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4w7kp"] Oct 05 07:14:07 crc kubenswrapper[4935]: W1005 07:14:07.841321 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2def690_05d7_4196_b383_58669b274802.slice/crio-6f46e40fc80c82db8bae92c1e702d601fc61f9b770e91662d88f508991957ad2 WatchSource:0}: Error finding container 6f46e40fc80c82db8bae92c1e702d601fc61f9b770e91662d88f508991957ad2: Status 404 returned error can't find the container with id 6f46e40fc80c82db8bae92c1e702d601fc61f9b770e91662d88f508991957ad2 Oct 05 07:14:07 crc kubenswrapper[4935]: I1005 07:14:07.996390 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4w7kp" event={"ID":"c2def690-05d7-4196-b383-58669b274802","Type":"ContainerStarted","Data":"6f46e40fc80c82db8bae92c1e702d601fc61f9b770e91662d88f508991957ad2"} Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.013192 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vlqf6"] Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.014455 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.017175 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.017397 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.029841 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vlqf6"] Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.059459 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.076733 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-5qg9s"] Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.137193 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-scripts\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.137608 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hhzj\" (UniqueName: \"kubernetes.io/projected/234e0c62-014f-40c5-834f-e897b9593cdb-kube-api-access-7hhzj\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.137629 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-config-data\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.137673 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.238861 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.238970 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-scripts\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.239060 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hhzj\" (UniqueName: \"kubernetes.io/projected/234e0c62-014f-40c5-834f-e897b9593cdb-kube-api-access-7hhzj\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.239082 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-config-data\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.240564 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.247652 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.260240 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-scripts\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.260808 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-config-data\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.263675 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.267301 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hhzj\" (UniqueName: \"kubernetes.io/projected/234e0c62-014f-40c5-834f-e897b9593cdb-kube-api-access-7hhzj\") pod \"nova-cell1-conductor-db-sync-vlqf6\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.342779 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.451909 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:08 crc kubenswrapper[4935]: I1005 07:14:08.816183 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vlqf6"] Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.022389 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4w7kp" event={"ID":"c2def690-05d7-4196-b383-58669b274802","Type":"ContainerStarted","Data":"b79d9e47fed04b770cc464950c913e995fbd2b7802e83c46a8abf39fd1d704d0"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.031977 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b51b1cb3-ea91-4535-833b-03138c9c9eb4","Type":"ContainerStarted","Data":"4dee96877d32bbbf2ac4bf9622abdb200eca452c03ea5c3ee6fd4a44adeb66b9"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.057604 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59688ed9-ff1f-4c93-961c-247789e17af0","Type":"ContainerStarted","Data":"3eabe4542012450e410ecf9e94b6a57a1d56b39eeae5f6390dd451a466374d1f"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.068342 4935 generic.go:334] "Generic (PLEG): container finished" podID="a02edeb1-77e8-4884-a922-3dc7d608b587" containerID="95737c65c8ee191ece2d2590b4f6194158f5c08fb6949f2e857aa29e0037e4c3" exitCode=0 Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.068642 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" event={"ID":"a02edeb1-77e8-4884-a922-3dc7d608b587","Type":"ContainerDied","Data":"95737c65c8ee191ece2d2590b4f6194158f5c08fb6949f2e857aa29e0037e4c3"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.068728 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" event={"ID":"a02edeb1-77e8-4884-a922-3dc7d608b587","Type":"ContainerStarted","Data":"bebcce117c4b9d97cca2915d750877ff31291775637b77a44b0d41b150a80cb9"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.077129 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-4w7kp" podStartSLOduration=3.077109247 podStartE2EDuration="3.077109247s" podCreationTimestamp="2025-10-05 07:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:09.041152806 +0000 UTC m=+1282.923779266" watchObservedRunningTime="2025-10-05 07:14:09.077109247 +0000 UTC m=+1282.959735707" Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.097509 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" event={"ID":"234e0c62-014f-40c5-834f-e897b9593cdb","Type":"ContainerStarted","Data":"1458d1b7cdd99956b967eb1f9853c2bd78c8f640dd1df44deaebf07a076374dd"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.097907 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" event={"ID":"234e0c62-014f-40c5-834f-e897b9593cdb","Type":"ContainerStarted","Data":"70426f1a131cfb7b516ceb7f5f83de91a18250bb5b4def3184819616b6d858b6"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.101008 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e","Type":"ContainerStarted","Data":"3824cff43b8b7dffeb3caeefa0800f5c4d093949183cd04f9801f7534c259421"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.106208 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609","Type":"ContainerStarted","Data":"aa30d25ef443bb4b0187bfec034855c982bdcc5756b84fdaf747c69c222da450"} Oct 05 07:14:09 crc kubenswrapper[4935]: I1005 07:14:09.130012 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" podStartSLOduration=2.129993106 podStartE2EDuration="2.129993106s" podCreationTimestamp="2025-10-05 07:14:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:09.116216232 +0000 UTC m=+1282.998842702" watchObservedRunningTime="2025-10-05 07:14:09.129993106 +0000 UTC m=+1283.012619566" Oct 05 07:14:10 crc kubenswrapper[4935]: I1005 07:14:10.118659 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" event={"ID":"a02edeb1-77e8-4884-a922-3dc7d608b587","Type":"ContainerStarted","Data":"9dd6b516cda99fd95d5c309707235a54553d259bf1fba2f5808a58484cacac39"} Oct 05 07:14:10 crc kubenswrapper[4935]: I1005 07:14:10.119180 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:10 crc kubenswrapper[4935]: I1005 07:14:10.147215 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" podStartSLOduration=3.147169489 podStartE2EDuration="3.147169489s" podCreationTimestamp="2025-10-05 07:14:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:10.142418623 +0000 UTC m=+1284.025045083" watchObservedRunningTime="2025-10-05 07:14:10.147169489 +0000 UTC m=+1284.029795949" Oct 05 07:14:10 crc kubenswrapper[4935]: I1005 07:14:10.806169 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:10 crc kubenswrapper[4935]: I1005 07:14:10.819653 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.136398 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609","Type":"ContainerStarted","Data":"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f"} Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.137016 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609","Type":"ContainerStarted","Data":"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7"} Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.136681 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerName="nova-metadata-log" containerID="cri-o://9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7" gracePeriod=30 Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.137055 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerName="nova-metadata-metadata" containerID="cri-o://3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f" gracePeriod=30 Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.138574 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b51b1cb3-ea91-4535-833b-03138c9c9eb4","Type":"ContainerStarted","Data":"80a1520f9593cdfba93bb8feb709352bfbc9abfeaca158e5202ac2c16aaf6787"} Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.148592 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59688ed9-ff1f-4c93-961c-247789e17af0","Type":"ContainerStarted","Data":"1728029339abc70034c54155cfbbf6d41048ac00622ca12313d214f4e7ec9fc9"} Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.148646 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59688ed9-ff1f-4c93-961c-247789e17af0","Type":"ContainerStarted","Data":"c9f2b76c7a59a55d95e48d3ed77f8499e7423b0c3445cb1a374ce0c2b6af518e"} Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.151016 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e","Type":"ContainerStarted","Data":"13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e"} Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.151135 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e" gracePeriod=30 Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.169101 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.002481933 podStartE2EDuration="6.169079703s" podCreationTimestamp="2025-10-05 07:14:06 +0000 UTC" firstStartedPulling="2025-10-05 07:14:08.26573159 +0000 UTC m=+1282.148358050" lastFinishedPulling="2025-10-05 07:14:11.43232936 +0000 UTC m=+1285.314955820" observedRunningTime="2025-10-05 07:14:12.161212745 +0000 UTC m=+1286.043839205" watchObservedRunningTime="2025-10-05 07:14:12.169079703 +0000 UTC m=+1286.051706173" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.196997 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.07048768 podStartE2EDuration="6.196865568s" podCreationTimestamp="2025-10-05 07:14:06 +0000 UTC" firstStartedPulling="2025-10-05 07:14:08.277047719 +0000 UTC m=+1282.159674179" lastFinishedPulling="2025-10-05 07:14:11.403425597 +0000 UTC m=+1285.286052067" observedRunningTime="2025-10-05 07:14:12.181632735 +0000 UTC m=+1286.064259195" watchObservedRunningTime="2025-10-05 07:14:12.196865568 +0000 UTC m=+1286.079492028" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.221331 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.280725213 podStartE2EDuration="5.221306635s" podCreationTimestamp="2025-10-05 07:14:07 +0000 UTC" firstStartedPulling="2025-10-05 07:14:08.463533263 +0000 UTC m=+1282.346159723" lastFinishedPulling="2025-10-05 07:14:11.404114675 +0000 UTC m=+1285.286741145" observedRunningTime="2025-10-05 07:14:12.198152772 +0000 UTC m=+1286.080779242" watchObservedRunningTime="2025-10-05 07:14:12.221306635 +0000 UTC m=+1286.103933095" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.226096 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.8962587210000001 podStartE2EDuration="5.226085701s" podCreationTimestamp="2025-10-05 07:14:07 +0000 UTC" firstStartedPulling="2025-10-05 07:14:08.073800362 +0000 UTC m=+1281.956426822" lastFinishedPulling="2025-10-05 07:14:11.403627342 +0000 UTC m=+1285.286253802" observedRunningTime="2025-10-05 07:14:12.216828316 +0000 UTC m=+1286.099454786" watchObservedRunningTime="2025-10-05 07:14:12.226085701 +0000 UTC m=+1286.108712151" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.504980 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.505373 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.574742 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.748451 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.769022 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.853074 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-logs\") pod \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.853228 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-combined-ca-bundle\") pod \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.853366 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64tgc\" (UniqueName: \"kubernetes.io/projected/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-kube-api-access-64tgc\") pod \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.853426 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-config-data\") pod \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\" (UID: \"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609\") " Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.853453 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-logs" (OuterVolumeSpecName: "logs") pod "b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" (UID: "b3b7a8ce-afa2-43b0-b53b-02c7b9b76609"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.853851 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.863259 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-kube-api-access-64tgc" (OuterVolumeSpecName: "kube-api-access-64tgc") pod "b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" (UID: "b3b7a8ce-afa2-43b0-b53b-02c7b9b76609"). InnerVolumeSpecName "kube-api-access-64tgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.899673 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" (UID: "b3b7a8ce-afa2-43b0-b53b-02c7b9b76609"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.909063 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-config-data" (OuterVolumeSpecName: "config-data") pod "b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" (UID: "b3b7a8ce-afa2-43b0-b53b-02c7b9b76609"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.957024 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.957060 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:12 crc kubenswrapper[4935]: I1005 07:14:12.957074 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64tgc\" (UniqueName: \"kubernetes.io/projected/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609-kube-api-access-64tgc\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.167817 4935 generic.go:334] "Generic (PLEG): container finished" podID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerID="3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f" exitCode=0 Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.168022 4935 generic.go:334] "Generic (PLEG): container finished" podID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerID="9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7" exitCode=143 Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.169154 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.172085 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609","Type":"ContainerDied","Data":"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f"} Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.172158 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609","Type":"ContainerDied","Data":"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7"} Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.172171 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3b7a8ce-afa2-43b0-b53b-02c7b9b76609","Type":"ContainerDied","Data":"aa30d25ef443bb4b0187bfec034855c982bdcc5756b84fdaf747c69c222da450"} Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.172191 4935 scope.go:117] "RemoveContainer" containerID="3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.216117 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.222912 4935 scope.go:117] "RemoveContainer" containerID="9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.226524 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.245257 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:13 crc kubenswrapper[4935]: E1005 07:14:13.245714 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerName="nova-metadata-metadata" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.245735 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerName="nova-metadata-metadata" Oct 05 07:14:13 crc kubenswrapper[4935]: E1005 07:14:13.245764 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerName="nova-metadata-log" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.245774 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerName="nova-metadata-log" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.246023 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerName="nova-metadata-log" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.246048 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" containerName="nova-metadata-metadata" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.247271 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.256514 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.256863 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.264939 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.265075 4935 scope.go:117] "RemoveContainer" containerID="3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f" Oct 05 07:14:13 crc kubenswrapper[4935]: E1005 07:14:13.266478 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f\": container with ID starting with 3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f not found: ID does not exist" containerID="3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.266514 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f"} err="failed to get container status \"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f\": rpc error: code = NotFound desc = could not find container \"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f\": container with ID starting with 3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f not found: ID does not exist" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.266537 4935 scope.go:117] "RemoveContainer" containerID="9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7" Oct 05 07:14:13 crc kubenswrapper[4935]: E1005 07:14:13.267017 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7\": container with ID starting with 9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7 not found: ID does not exist" containerID="9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.267041 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7"} err="failed to get container status \"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7\": rpc error: code = NotFound desc = could not find container \"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7\": container with ID starting with 9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7 not found: ID does not exist" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.267056 4935 scope.go:117] "RemoveContainer" containerID="3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.267261 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f"} err="failed to get container status \"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f\": rpc error: code = NotFound desc = could not find container \"3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f\": container with ID starting with 3af37fdf56586b5a780a500d2fa996642dac83f334a179a01a46bda097f8817f not found: ID does not exist" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.267280 4935 scope.go:117] "RemoveContainer" containerID="9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.267587 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7"} err="failed to get container status \"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7\": rpc error: code = NotFound desc = could not find container \"9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7\": container with ID starting with 9a068c359e34ebd2f3ab00ace259f63deef29a4803dc4dba05fcb31c5b4665c7 not found: ID does not exist" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.365968 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-config-data\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.366025 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc6zk\" (UniqueName: \"kubernetes.io/projected/760e837b-9d99-4b46-b176-220b5ef8b5bb-kube-api-access-dc6zk\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.366088 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.366141 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/760e837b-9d99-4b46-b176-220b5ef8b5bb-logs\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.366180 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.468458 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/760e837b-9d99-4b46-b176-220b5ef8b5bb-logs\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.468582 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.468672 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-config-data\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.468730 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc6zk\" (UniqueName: \"kubernetes.io/projected/760e837b-9d99-4b46-b176-220b5ef8b5bb-kube-api-access-dc6zk\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.468834 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.470602 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/760e837b-9d99-4b46-b176-220b5ef8b5bb-logs\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.474650 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.475790 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-config-data\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.486116 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.498425 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc6zk\" (UniqueName: \"kubernetes.io/projected/760e837b-9d99-4b46-b176-220b5ef8b5bb-kube-api-access-dc6zk\") pod \"nova-metadata-0\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " pod="openstack/nova-metadata-0" Oct 05 07:14:13 crc kubenswrapper[4935]: I1005 07:14:13.602019 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:14 crc kubenswrapper[4935]: I1005 07:14:14.164260 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:14 crc kubenswrapper[4935]: W1005 07:14:14.164446 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod760e837b_9d99_4b46_b176_220b5ef8b5bb.slice/crio-36f4f7ad37912683e851561b85e07166973459ea136ee06f00ee11839c1dfa2c WatchSource:0}: Error finding container 36f4f7ad37912683e851561b85e07166973459ea136ee06f00ee11839c1dfa2c: Status 404 returned error can't find the container with id 36f4f7ad37912683e851561b85e07166973459ea136ee06f00ee11839c1dfa2c Oct 05 07:14:14 crc kubenswrapper[4935]: I1005 07:14:14.189369 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"760e837b-9d99-4b46-b176-220b5ef8b5bb","Type":"ContainerStarted","Data":"36f4f7ad37912683e851561b85e07166973459ea136ee06f00ee11839c1dfa2c"} Oct 05 07:14:14 crc kubenswrapper[4935]: I1005 07:14:14.289726 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:14:14 crc kubenswrapper[4935]: I1005 07:14:14.290104 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:14:14 crc kubenswrapper[4935]: I1005 07:14:14.789277 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3b7a8ce-afa2-43b0-b53b-02c7b9b76609" path="/var/lib/kubelet/pods/b3b7a8ce-afa2-43b0-b53b-02c7b9b76609/volumes" Oct 05 07:14:15 crc kubenswrapper[4935]: I1005 07:14:15.205191 4935 generic.go:334] "Generic (PLEG): container finished" podID="234e0c62-014f-40c5-834f-e897b9593cdb" containerID="1458d1b7cdd99956b967eb1f9853c2bd78c8f640dd1df44deaebf07a076374dd" exitCode=0 Oct 05 07:14:15 crc kubenswrapper[4935]: I1005 07:14:15.205312 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" event={"ID":"234e0c62-014f-40c5-834f-e897b9593cdb","Type":"ContainerDied","Data":"1458d1b7cdd99956b967eb1f9853c2bd78c8f640dd1df44deaebf07a076374dd"} Oct 05 07:14:15 crc kubenswrapper[4935]: I1005 07:14:15.208752 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"760e837b-9d99-4b46-b176-220b5ef8b5bb","Type":"ContainerStarted","Data":"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35"} Oct 05 07:14:15 crc kubenswrapper[4935]: I1005 07:14:15.208809 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"760e837b-9d99-4b46-b176-220b5ef8b5bb","Type":"ContainerStarted","Data":"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6"} Oct 05 07:14:15 crc kubenswrapper[4935]: I1005 07:14:15.257123 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.257101074 podStartE2EDuration="2.257101074s" podCreationTimestamp="2025-10-05 07:14:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:15.257020892 +0000 UTC m=+1289.139647392" watchObservedRunningTime="2025-10-05 07:14:15.257101074 +0000 UTC m=+1289.139727554" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.222789 4935 generic.go:334] "Generic (PLEG): container finished" podID="c2def690-05d7-4196-b383-58669b274802" containerID="b79d9e47fed04b770cc464950c913e995fbd2b7802e83c46a8abf39fd1d704d0" exitCode=0 Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.222904 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4w7kp" event={"ID":"c2def690-05d7-4196-b383-58669b274802","Type":"ContainerDied","Data":"b79d9e47fed04b770cc464950c913e995fbd2b7802e83c46a8abf39fd1d704d0"} Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.606142 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.743334 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-config-data\") pod \"234e0c62-014f-40c5-834f-e897b9593cdb\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.743486 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-scripts\") pod \"234e0c62-014f-40c5-834f-e897b9593cdb\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.743543 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hhzj\" (UniqueName: \"kubernetes.io/projected/234e0c62-014f-40c5-834f-e897b9593cdb-kube-api-access-7hhzj\") pod \"234e0c62-014f-40c5-834f-e897b9593cdb\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.743626 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-combined-ca-bundle\") pod \"234e0c62-014f-40c5-834f-e897b9593cdb\" (UID: \"234e0c62-014f-40c5-834f-e897b9593cdb\") " Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.750405 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-scripts" (OuterVolumeSpecName: "scripts") pod "234e0c62-014f-40c5-834f-e897b9593cdb" (UID: "234e0c62-014f-40c5-834f-e897b9593cdb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.750652 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234e0c62-014f-40c5-834f-e897b9593cdb-kube-api-access-7hhzj" (OuterVolumeSpecName: "kube-api-access-7hhzj") pod "234e0c62-014f-40c5-834f-e897b9593cdb" (UID: "234e0c62-014f-40c5-834f-e897b9593cdb"). InnerVolumeSpecName "kube-api-access-7hhzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.773840 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-config-data" (OuterVolumeSpecName: "config-data") pod "234e0c62-014f-40c5-834f-e897b9593cdb" (UID: "234e0c62-014f-40c5-834f-e897b9593cdb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.785670 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "234e0c62-014f-40c5-834f-e897b9593cdb" (UID: "234e0c62-014f-40c5-834f-e897b9593cdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.847059 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.847210 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.847286 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/234e0c62-014f-40c5-834f-e897b9593cdb-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:16 crc kubenswrapper[4935]: I1005 07:14:16.847367 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hhzj\" (UniqueName: \"kubernetes.io/projected/234e0c62-014f-40c5-834f-e897b9593cdb-kube-api-access-7hhzj\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.233764 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" event={"ID":"234e0c62-014f-40c5-834f-e897b9593cdb","Type":"ContainerDied","Data":"70426f1a131cfb7b516ceb7f5f83de91a18250bb5b4def3184819616b6d858b6"} Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.234212 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70426f1a131cfb7b516ceb7f5f83de91a18250bb5b4def3184819616b6d858b6" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.233967 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vlqf6" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.330467 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:14:17 crc kubenswrapper[4935]: E1005 07:14:17.336508 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234e0c62-014f-40c5-834f-e897b9593cdb" containerName="nova-cell1-conductor-db-sync" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.336627 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="234e0c62-014f-40c5-834f-e897b9593cdb" containerName="nova-cell1-conductor-db-sync" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.338979 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="234e0c62-014f-40c5-834f-e897b9593cdb" containerName="nova-cell1-conductor-db-sync" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.339827 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.343772 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.346542 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.419143 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.430196 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.430270 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.463261 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.463312 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.463452 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcdqm\" (UniqueName: \"kubernetes.io/projected/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-kube-api-access-pcdqm\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.485759 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-k9hln"] Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.486211 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" podUID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" containerName="dnsmasq-dns" containerID="cri-o://db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92" gracePeriod=10 Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.566046 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcdqm\" (UniqueName: \"kubernetes.io/projected/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-kube-api-access-pcdqm\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.566496 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.566537 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.573132 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.574212 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.578805 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.586525 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcdqm\" (UniqueName: \"kubernetes.io/projected/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-kube-api-access-pcdqm\") pod \"nova-cell1-conductor-0\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.616128 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.688407 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.712225 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.894672 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-combined-ca-bundle\") pod \"c2def690-05d7-4196-b383-58669b274802\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.895011 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-scripts\") pod \"c2def690-05d7-4196-b383-58669b274802\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.895054 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-config-data\") pod \"c2def690-05d7-4196-b383-58669b274802\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.895157 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj9q6\" (UniqueName: \"kubernetes.io/projected/c2def690-05d7-4196-b383-58669b274802-kube-api-access-wj9q6\") pod \"c2def690-05d7-4196-b383-58669b274802\" (UID: \"c2def690-05d7-4196-b383-58669b274802\") " Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.901071 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-scripts" (OuterVolumeSpecName: "scripts") pod "c2def690-05d7-4196-b383-58669b274802" (UID: "c2def690-05d7-4196-b383-58669b274802"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.905249 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2def690-05d7-4196-b383-58669b274802-kube-api-access-wj9q6" (OuterVolumeSpecName: "kube-api-access-wj9q6") pod "c2def690-05d7-4196-b383-58669b274802" (UID: "c2def690-05d7-4196-b383-58669b274802"). InnerVolumeSpecName "kube-api-access-wj9q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.929446 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-config-data" (OuterVolumeSpecName: "config-data") pod "c2def690-05d7-4196-b383-58669b274802" (UID: "c2def690-05d7-4196-b383-58669b274802"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.929731 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2def690-05d7-4196-b383-58669b274802" (UID: "c2def690-05d7-4196-b383-58669b274802"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.997627 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj9q6\" (UniqueName: \"kubernetes.io/projected/c2def690-05d7-4196-b383-58669b274802-kube-api-access-wj9q6\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.997661 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.997673 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.997685 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2def690-05d7-4196-b383-58669b274802-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:17 crc kubenswrapper[4935]: I1005 07:14:17.998551 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.099199 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-swift-storage-0\") pod \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.099259 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-config\") pod \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.099289 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-nb\") pod \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.100049 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvfdb\" (UniqueName: \"kubernetes.io/projected/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-kube-api-access-cvfdb\") pod \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.100117 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-svc\") pod \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.100232 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-sb\") pod \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\" (UID: \"eb4c0762-88dd-404e-8cfc-6ed3860f41b3\") " Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.103126 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-kube-api-access-cvfdb" (OuterVolumeSpecName: "kube-api-access-cvfdb") pod "eb4c0762-88dd-404e-8cfc-6ed3860f41b3" (UID: "eb4c0762-88dd-404e-8cfc-6ed3860f41b3"). InnerVolumeSpecName "kube-api-access-cvfdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.148284 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb4c0762-88dd-404e-8cfc-6ed3860f41b3" (UID: "eb4c0762-88dd-404e-8cfc-6ed3860f41b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.150928 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-config" (OuterVolumeSpecName: "config") pod "eb4c0762-88dd-404e-8cfc-6ed3860f41b3" (UID: "eb4c0762-88dd-404e-8cfc-6ed3860f41b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.153980 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eb4c0762-88dd-404e-8cfc-6ed3860f41b3" (UID: "eb4c0762-88dd-404e-8cfc-6ed3860f41b3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.160406 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb4c0762-88dd-404e-8cfc-6ed3860f41b3" (UID: "eb4c0762-88dd-404e-8cfc-6ed3860f41b3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.163447 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb4c0762-88dd-404e-8cfc-6ed3860f41b3" (UID: "eb4c0762-88dd-404e-8cfc-6ed3860f41b3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.202591 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.202623 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.202633 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.202641 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.202651 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvfdb\" (UniqueName: \"kubernetes.io/projected/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-kube-api-access-cvfdb\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.202660 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb4c0762-88dd-404e-8cfc-6ed3860f41b3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.230322 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:14:18 crc kubenswrapper[4935]: W1005 07:14:18.232089 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77853c76_2bbe_42e5_a0ef_03d5c5fe2402.slice/crio-cd8f2c98f54599f7e46b0b16cd909e8d1b6113aa2db919860d5ca310abd5790f WatchSource:0}: Error finding container cd8f2c98f54599f7e46b0b16cd909e8d1b6113aa2db919860d5ca310abd5790f: Status 404 returned error can't find the container with id cd8f2c98f54599f7e46b0b16cd909e8d1b6113aa2db919860d5ca310abd5790f Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.243291 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"77853c76-2bbe-42e5-a0ef-03d5c5fe2402","Type":"ContainerStarted","Data":"cd8f2c98f54599f7e46b0b16cd909e8d1b6113aa2db919860d5ca310abd5790f"} Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.246692 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4w7kp" event={"ID":"c2def690-05d7-4196-b383-58669b274802","Type":"ContainerDied","Data":"6f46e40fc80c82db8bae92c1e702d601fc61f9b770e91662d88f508991957ad2"} Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.246765 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f46e40fc80c82db8bae92c1e702d601fc61f9b770e91662d88f508991957ad2" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.246757 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4w7kp" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.249968 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.249993 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" event={"ID":"eb4c0762-88dd-404e-8cfc-6ed3860f41b3","Type":"ContainerDied","Data":"db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92"} Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.250048 4935 scope.go:117] "RemoveContainer" containerID="db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.258258 4935 generic.go:334] "Generic (PLEG): container finished" podID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" containerID="db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92" exitCode=0 Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.258555 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-k9hln" event={"ID":"eb4c0762-88dd-404e-8cfc-6ed3860f41b3","Type":"ContainerDied","Data":"26cf71adea4779a7058bfd6cc8b7a2691c2b9ea0046d84682731c6c608f3c4c7"} Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.307488 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.313798 4935 scope.go:117] "RemoveContainer" containerID="717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.316499 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-k9hln"] Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.327312 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-k9hln"] Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.343024 4935 scope.go:117] "RemoveContainer" containerID="db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92" Oct 05 07:14:18 crc kubenswrapper[4935]: E1005 07:14:18.343448 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92\": container with ID starting with db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92 not found: ID does not exist" containerID="db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.343478 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92"} err="failed to get container status \"db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92\": rpc error: code = NotFound desc = could not find container \"db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92\": container with ID starting with db66229fb91b2d2daccd651d57dab47f7f47927c6ebcca5eda3db81616eced92 not found: ID does not exist" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.343499 4935 scope.go:117] "RemoveContainer" containerID="717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab" Oct 05 07:14:18 crc kubenswrapper[4935]: E1005 07:14:18.343698 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab\": container with ID starting with 717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab not found: ID does not exist" containerID="717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.343719 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab"} err="failed to get container status \"717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab\": rpc error: code = NotFound desc = could not find container \"717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab\": container with ID starting with 717128134f44aeee174b59a6ef26a10474ff2f7eddb18f92722db821da85aeab not found: ID does not exist" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.427449 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.427698 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-log" containerID="cri-o://c9f2b76c7a59a55d95e48d3ed77f8499e7423b0c3445cb1a374ce0c2b6af518e" gracePeriod=30 Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.428102 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-api" containerID="cri-o://1728029339abc70034c54155cfbbf6d41048ac00622ca12313d214f4e7ec9fc9" gracePeriod=30 Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.432197 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.432481 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": EOF (Client.Timeout exceeded while awaiting headers)" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.443770 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.444708 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerName="nova-metadata-metadata" containerID="cri-o://1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35" gracePeriod=30 Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.444387 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerName="nova-metadata-log" containerID="cri-o://83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6" gracePeriod=30 Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.602516 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.602584 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.737100 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:18 crc kubenswrapper[4935]: I1005 07:14:18.818570 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" path="/var/lib/kubelet/pods/eb4c0762-88dd-404e-8cfc-6ed3860f41b3/volumes" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.039629 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.124374 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-config-data\") pod \"760e837b-9d99-4b46-b176-220b5ef8b5bb\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.124417 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-nova-metadata-tls-certs\") pod \"760e837b-9d99-4b46-b176-220b5ef8b5bb\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.124472 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/760e837b-9d99-4b46-b176-220b5ef8b5bb-logs\") pod \"760e837b-9d99-4b46-b176-220b5ef8b5bb\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.124690 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc6zk\" (UniqueName: \"kubernetes.io/projected/760e837b-9d99-4b46-b176-220b5ef8b5bb-kube-api-access-dc6zk\") pod \"760e837b-9d99-4b46-b176-220b5ef8b5bb\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.124761 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-combined-ca-bundle\") pod \"760e837b-9d99-4b46-b176-220b5ef8b5bb\" (UID: \"760e837b-9d99-4b46-b176-220b5ef8b5bb\") " Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.135090 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/760e837b-9d99-4b46-b176-220b5ef8b5bb-logs" (OuterVolumeSpecName: "logs") pod "760e837b-9d99-4b46-b176-220b5ef8b5bb" (UID: "760e837b-9d99-4b46-b176-220b5ef8b5bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.154690 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/760e837b-9d99-4b46-b176-220b5ef8b5bb-kube-api-access-dc6zk" (OuterVolumeSpecName: "kube-api-access-dc6zk") pod "760e837b-9d99-4b46-b176-220b5ef8b5bb" (UID: "760e837b-9d99-4b46-b176-220b5ef8b5bb"). InnerVolumeSpecName "kube-api-access-dc6zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.173292 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-config-data" (OuterVolumeSpecName: "config-data") pod "760e837b-9d99-4b46-b176-220b5ef8b5bb" (UID: "760e837b-9d99-4b46-b176-220b5ef8b5bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.182757 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "760e837b-9d99-4b46-b176-220b5ef8b5bb" (UID: "760e837b-9d99-4b46-b176-220b5ef8b5bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.203042 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "760e837b-9d99-4b46-b176-220b5ef8b5bb" (UID: "760e837b-9d99-4b46-b176-220b5ef8b5bb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.226642 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc6zk\" (UniqueName: \"kubernetes.io/projected/760e837b-9d99-4b46-b176-220b5ef8b5bb-kube-api-access-dc6zk\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.226671 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.226681 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.226689 4935 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/760e837b-9d99-4b46-b176-220b5ef8b5bb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.226698 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/760e837b-9d99-4b46-b176-220b5ef8b5bb-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.272319 4935 generic.go:334] "Generic (PLEG): container finished" podID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerID="1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35" exitCode=0 Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.272352 4935 generic.go:334] "Generic (PLEG): container finished" podID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerID="83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6" exitCode=143 Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.272390 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"760e837b-9d99-4b46-b176-220b5ef8b5bb","Type":"ContainerDied","Data":"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35"} Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.272415 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"760e837b-9d99-4b46-b176-220b5ef8b5bb","Type":"ContainerDied","Data":"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6"} Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.272425 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"760e837b-9d99-4b46-b176-220b5ef8b5bb","Type":"ContainerDied","Data":"36f4f7ad37912683e851561b85e07166973459ea136ee06f00ee11839c1dfa2c"} Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.272440 4935 scope.go:117] "RemoveContainer" containerID="1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.272512 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.291262 4935 generic.go:334] "Generic (PLEG): container finished" podID="59688ed9-ff1f-4c93-961c-247789e17af0" containerID="c9f2b76c7a59a55d95e48d3ed77f8499e7423b0c3445cb1a374ce0c2b6af518e" exitCode=143 Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.291367 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59688ed9-ff1f-4c93-961c-247789e17af0","Type":"ContainerDied","Data":"c9f2b76c7a59a55d95e48d3ed77f8499e7423b0c3445cb1a374ce0c2b6af518e"} Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.309441 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"77853c76-2bbe-42e5-a0ef-03d5c5fe2402","Type":"ContainerStarted","Data":"f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465"} Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.309492 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.336925 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.337879 4935 scope.go:117] "RemoveContainer" containerID="83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.348114 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.367974 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:19 crc kubenswrapper[4935]: E1005 07:14:19.368493 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2def690-05d7-4196-b383-58669b274802" containerName="nova-manage" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368518 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2def690-05d7-4196-b383-58669b274802" containerName="nova-manage" Oct 05 07:14:19 crc kubenswrapper[4935]: E1005 07:14:19.368555 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" containerName="dnsmasq-dns" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368565 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" containerName="dnsmasq-dns" Oct 05 07:14:19 crc kubenswrapper[4935]: E1005 07:14:19.368580 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" containerName="init" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368590 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" containerName="init" Oct 05 07:14:19 crc kubenswrapper[4935]: E1005 07:14:19.368608 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerName="nova-metadata-log" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368616 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerName="nova-metadata-log" Oct 05 07:14:19 crc kubenswrapper[4935]: E1005 07:14:19.368639 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerName="nova-metadata-metadata" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368647 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerName="nova-metadata-metadata" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368853 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerName="nova-metadata-metadata" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368874 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb4c0762-88dd-404e-8cfc-6ed3860f41b3" containerName="dnsmasq-dns" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368910 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" containerName="nova-metadata-log" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.368923 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2def690-05d7-4196-b383-58669b274802" containerName="nova-manage" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.370149 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.377489 4935 scope.go:117] "RemoveContainer" containerID="1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.377734 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.378000 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 07:14:19 crc kubenswrapper[4935]: E1005 07:14:19.379660 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35\": container with ID starting with 1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35 not found: ID does not exist" containerID="1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.379734 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35"} err="failed to get container status \"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35\": rpc error: code = NotFound desc = could not find container \"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35\": container with ID starting with 1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35 not found: ID does not exist" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.379766 4935 scope.go:117] "RemoveContainer" containerID="83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6" Oct 05 07:14:19 crc kubenswrapper[4935]: E1005 07:14:19.380170 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6\": container with ID starting with 83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6 not found: ID does not exist" containerID="83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.380196 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6"} err="failed to get container status \"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6\": rpc error: code = NotFound desc = could not find container \"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6\": container with ID starting with 83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6 not found: ID does not exist" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.380211 4935 scope.go:117] "RemoveContainer" containerID="1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.381083 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35"} err="failed to get container status \"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35\": rpc error: code = NotFound desc = could not find container \"1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35\": container with ID starting with 1c7e8a3d4b940eee116c6afcda70de30a71c44bd6b3393aa9261572ac01a0a35 not found: ID does not exist" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.381104 4935 scope.go:117] "RemoveContainer" containerID="83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.381344 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6"} err="failed to get container status \"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6\": rpc error: code = NotFound desc = could not find container \"83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6\": container with ID starting with 83a3f86e5c60c99fa57710450ca2a0d6937a736b5a2aba82e4a28702d7baa5c6 not found: ID does not exist" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.400962 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.40093899 podStartE2EDuration="2.40093899s" podCreationTimestamp="2025-10-05 07:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:19.32609056 +0000 UTC m=+1293.208717020" watchObservedRunningTime="2025-10-05 07:14:19.40093899 +0000 UTC m=+1293.283565450" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.410607 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.428991 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dpz5\" (UniqueName: \"kubernetes.io/projected/9400b67d-c149-43bb-b020-e3d2d93f32c1-kube-api-access-7dpz5\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.429098 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-config-data\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.429126 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9400b67d-c149-43bb-b020-e3d2d93f32c1-logs\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.429155 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.429171 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.530446 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-config-data\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.530508 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9400b67d-c149-43bb-b020-e3d2d93f32c1-logs\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.530553 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.530578 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.530686 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dpz5\" (UniqueName: \"kubernetes.io/projected/9400b67d-c149-43bb-b020-e3d2d93f32c1-kube-api-access-7dpz5\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.531914 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9400b67d-c149-43bb-b020-e3d2d93f32c1-logs\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.535577 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.535603 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.536424 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-config-data\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.550460 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dpz5\" (UniqueName: \"kubernetes.io/projected/9400b67d-c149-43bb-b020-e3d2d93f32c1-kube-api-access-7dpz5\") pod \"nova-metadata-0\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " pod="openstack/nova-metadata-0" Oct 05 07:14:19 crc kubenswrapper[4935]: I1005 07:14:19.692013 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:14:20 crc kubenswrapper[4935]: I1005 07:14:20.213648 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:14:20 crc kubenswrapper[4935]: I1005 07:14:20.325076 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9400b67d-c149-43bb-b020-e3d2d93f32c1","Type":"ContainerStarted","Data":"a0acc354536afa4c89bc273bbba8dc40549cd0d35e89d7f3ad2dbae96123e18f"} Oct 05 07:14:20 crc kubenswrapper[4935]: I1005 07:14:20.325304 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b51b1cb3-ea91-4535-833b-03138c9c9eb4" containerName="nova-scheduler-scheduler" containerID="cri-o://80a1520f9593cdfba93bb8feb709352bfbc9abfeaca158e5202ac2c16aaf6787" gracePeriod=30 Oct 05 07:14:20 crc kubenswrapper[4935]: I1005 07:14:20.799239 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="760e837b-9d99-4b46-b176-220b5ef8b5bb" path="/var/lib/kubelet/pods/760e837b-9d99-4b46-b176-220b5ef8b5bb/volumes" Oct 05 07:14:21 crc kubenswrapper[4935]: I1005 07:14:21.335553 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9400b67d-c149-43bb-b020-e3d2d93f32c1","Type":"ContainerStarted","Data":"8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97"} Oct 05 07:14:21 crc kubenswrapper[4935]: I1005 07:14:21.335855 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9400b67d-c149-43bb-b020-e3d2d93f32c1","Type":"ContainerStarted","Data":"a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a"} Oct 05 07:14:21 crc kubenswrapper[4935]: I1005 07:14:21.373312 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.373289184 podStartE2EDuration="2.373289184s" podCreationTimestamp="2025-10-05 07:14:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:21.365493028 +0000 UTC m=+1295.248119498" watchObservedRunningTime="2025-10-05 07:14:21.373289184 +0000 UTC m=+1295.255915654" Oct 05 07:14:22 crc kubenswrapper[4935]: E1005 07:14:22.576662 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80a1520f9593cdfba93bb8feb709352bfbc9abfeaca158e5202ac2c16aaf6787" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:14:22 crc kubenswrapper[4935]: E1005 07:14:22.578842 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80a1520f9593cdfba93bb8feb709352bfbc9abfeaca158e5202ac2c16aaf6787" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:14:22 crc kubenswrapper[4935]: E1005 07:14:22.580364 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80a1520f9593cdfba93bb8feb709352bfbc9abfeaca158e5202ac2c16aaf6787" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:14:22 crc kubenswrapper[4935]: E1005 07:14:22.580413 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b51b1cb3-ea91-4535-833b-03138c9c9eb4" containerName="nova-scheduler-scheduler" Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.359950 4935 generic.go:334] "Generic (PLEG): container finished" podID="b51b1cb3-ea91-4535-833b-03138c9c9eb4" containerID="80a1520f9593cdfba93bb8feb709352bfbc9abfeaca158e5202ac2c16aaf6787" exitCode=0 Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.360396 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b51b1cb3-ea91-4535-833b-03138c9c9eb4","Type":"ContainerDied","Data":"80a1520f9593cdfba93bb8feb709352bfbc9abfeaca158e5202ac2c16aaf6787"} Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.464323 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.620704 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chxcv\" (UniqueName: \"kubernetes.io/projected/b51b1cb3-ea91-4535-833b-03138c9c9eb4-kube-api-access-chxcv\") pod \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.620885 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-config-data\") pod \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.620941 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-combined-ca-bundle\") pod \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\" (UID: \"b51b1cb3-ea91-4535-833b-03138c9c9eb4\") " Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.629134 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51b1cb3-ea91-4535-833b-03138c9c9eb4-kube-api-access-chxcv" (OuterVolumeSpecName: "kube-api-access-chxcv") pod "b51b1cb3-ea91-4535-833b-03138c9c9eb4" (UID: "b51b1cb3-ea91-4535-833b-03138c9c9eb4"). InnerVolumeSpecName "kube-api-access-chxcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.647458 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-config-data" (OuterVolumeSpecName: "config-data") pod "b51b1cb3-ea91-4535-833b-03138c9c9eb4" (UID: "b51b1cb3-ea91-4535-833b-03138c9c9eb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.655490 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b51b1cb3-ea91-4535-833b-03138c9c9eb4" (UID: "b51b1cb3-ea91-4535-833b-03138c9c9eb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.724098 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.724149 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1cb3-ea91-4535-833b-03138c9c9eb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:23 crc kubenswrapper[4935]: I1005 07:14:23.724170 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chxcv\" (UniqueName: \"kubernetes.io/projected/b51b1cb3-ea91-4535-833b-03138c9c9eb4-kube-api-access-chxcv\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.373188 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b51b1cb3-ea91-4535-833b-03138c9c9eb4","Type":"ContainerDied","Data":"4dee96877d32bbbf2ac4bf9622abdb200eca452c03ea5c3ee6fd4a44adeb66b9"} Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.373244 4935 scope.go:117] "RemoveContainer" containerID="80a1520f9593cdfba93bb8feb709352bfbc9abfeaca158e5202ac2c16aaf6787" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.373279 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.375581 4935 generic.go:334] "Generic (PLEG): container finished" podID="59688ed9-ff1f-4c93-961c-247789e17af0" containerID="1728029339abc70034c54155cfbbf6d41048ac00622ca12313d214f4e7ec9fc9" exitCode=0 Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.375627 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59688ed9-ff1f-4c93-961c-247789e17af0","Type":"ContainerDied","Data":"1728029339abc70034c54155cfbbf6d41048ac00622ca12313d214f4e7ec9fc9"} Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.375657 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59688ed9-ff1f-4c93-961c-247789e17af0","Type":"ContainerDied","Data":"3eabe4542012450e410ecf9e94b6a57a1d56b39eeae5f6390dd451a466374d1f"} Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.375670 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3eabe4542012450e410ecf9e94b6a57a1d56b39eeae5f6390dd451a466374d1f" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.438322 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.453450 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.489394 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.520648 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:24 crc kubenswrapper[4935]: E1005 07:14:24.521193 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-log" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.521221 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-log" Oct 05 07:14:24 crc kubenswrapper[4935]: E1005 07:14:24.521274 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51b1cb3-ea91-4535-833b-03138c9c9eb4" containerName="nova-scheduler-scheduler" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.521284 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51b1cb3-ea91-4535-833b-03138c9c9eb4" containerName="nova-scheduler-scheduler" Oct 05 07:14:24 crc kubenswrapper[4935]: E1005 07:14:24.521307 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-api" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.521316 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-api" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.521586 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51b1cb3-ea91-4535-833b-03138c9c9eb4" containerName="nova-scheduler-scheduler" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.521609 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-api" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.521643 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" containerName="nova-api-log" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.522655 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.525053 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.539132 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59688ed9-ff1f-4c93-961c-247789e17af0-logs\") pod \"59688ed9-ff1f-4c93-961c-247789e17af0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.539248 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv5vp\" (UniqueName: \"kubernetes.io/projected/59688ed9-ff1f-4c93-961c-247789e17af0-kube-api-access-rv5vp\") pod \"59688ed9-ff1f-4c93-961c-247789e17af0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.539274 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-combined-ca-bundle\") pod \"59688ed9-ff1f-4c93-961c-247789e17af0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.539529 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-config-data\") pod \"59688ed9-ff1f-4c93-961c-247789e17af0\" (UID: \"59688ed9-ff1f-4c93-961c-247789e17af0\") " Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.539561 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59688ed9-ff1f-4c93-961c-247789e17af0-logs" (OuterVolumeSpecName: "logs") pod "59688ed9-ff1f-4c93-961c-247789e17af0" (UID: "59688ed9-ff1f-4c93-961c-247789e17af0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.539988 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59688ed9-ff1f-4c93-961c-247789e17af0-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.542449 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.544018 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59688ed9-ff1f-4c93-961c-247789e17af0-kube-api-access-rv5vp" (OuterVolumeSpecName: "kube-api-access-rv5vp") pod "59688ed9-ff1f-4c93-961c-247789e17af0" (UID: "59688ed9-ff1f-4c93-961c-247789e17af0"). InnerVolumeSpecName "kube-api-access-rv5vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.566626 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59688ed9-ff1f-4c93-961c-247789e17af0" (UID: "59688ed9-ff1f-4c93-961c-247789e17af0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.567720 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-config-data" (OuterVolumeSpecName: "config-data") pod "59688ed9-ff1f-4c93-961c-247789e17af0" (UID: "59688ed9-ff1f-4c93-961c-247789e17af0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.642700 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-config-data\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.642790 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6cfv\" (UniqueName: \"kubernetes.io/projected/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-kube-api-access-z6cfv\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.643166 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.643371 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.643390 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv5vp\" (UniqueName: \"kubernetes.io/projected/59688ed9-ff1f-4c93-961c-247789e17af0-kube-api-access-rv5vp\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.643403 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59688ed9-ff1f-4c93-961c-247789e17af0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.693116 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.693158 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.745305 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-config-data\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.745388 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6cfv\" (UniqueName: \"kubernetes.io/projected/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-kube-api-access-z6cfv\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.745471 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.749107 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.749181 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-config-data\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.760213 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6cfv\" (UniqueName: \"kubernetes.io/projected/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-kube-api-access-z6cfv\") pod \"nova-scheduler-0\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " pod="openstack/nova-scheduler-0" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.790623 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51b1cb3-ea91-4535-833b-03138c9c9eb4" path="/var/lib/kubelet/pods/b51b1cb3-ea91-4535-833b-03138c9c9eb4/volumes" Oct 05 07:14:24 crc kubenswrapper[4935]: I1005 07:14:24.842763 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.358999 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:14:25 crc kubenswrapper[4935]: W1005 07:14:25.367793 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fa41a7b_d580_42c7_b7ce_55f1799b3dab.slice/crio-872f15980ca731abcc238bb1cc2e15c97781a8822f5acb47756c7f7b124a43ce WatchSource:0}: Error finding container 872f15980ca731abcc238bb1cc2e15c97781a8822f5acb47756c7f7b124a43ce: Status 404 returned error can't find the container with id 872f15980ca731abcc238bb1cc2e15c97781a8822f5acb47756c7f7b124a43ce Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.386348 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5fa41a7b-d580-42c7-b7ce-55f1799b3dab","Type":"ContainerStarted","Data":"872f15980ca731abcc238bb1cc2e15c97781a8822f5acb47756c7f7b124a43ce"} Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.387625 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.512971 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.532446 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.539711 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.543520 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.548284 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.555347 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.670244 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cklk\" (UniqueName: \"kubernetes.io/projected/9006ed82-292d-4af7-8851-fe468ea6f19a-kube-api-access-4cklk\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.670493 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9006ed82-292d-4af7-8851-fe468ea6f19a-logs\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.670658 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-config-data\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.671505 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.773168 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.773246 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cklk\" (UniqueName: \"kubernetes.io/projected/9006ed82-292d-4af7-8851-fe468ea6f19a-kube-api-access-4cklk\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.773289 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9006ed82-292d-4af7-8851-fe468ea6f19a-logs\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.773307 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-config-data\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.774004 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9006ed82-292d-4af7-8851-fe468ea6f19a-logs\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.778723 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.782332 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-config-data\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.791845 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cklk\" (UniqueName: \"kubernetes.io/projected/9006ed82-292d-4af7-8851-fe468ea6f19a-kube-api-access-4cklk\") pod \"nova-api-0\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " pod="openstack/nova-api-0" Oct 05 07:14:25 crc kubenswrapper[4935]: I1005 07:14:25.868293 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:26 crc kubenswrapper[4935]: I1005 07:14:26.309477 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:26 crc kubenswrapper[4935]: I1005 07:14:26.402673 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5fa41a7b-d580-42c7-b7ce-55f1799b3dab","Type":"ContainerStarted","Data":"00055eed91603d646bb93620f446eb6fdaa145639403015e47a6633ffda568ee"} Oct 05 07:14:26 crc kubenswrapper[4935]: I1005 07:14:26.407362 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9006ed82-292d-4af7-8851-fe468ea6f19a","Type":"ContainerStarted","Data":"877a2a14723c5e8aefa38d878377ad7723232f520dfcbca833472a4560d39fe7"} Oct 05 07:14:26 crc kubenswrapper[4935]: I1005 07:14:26.424246 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.424228389 podStartE2EDuration="2.424228389s" podCreationTimestamp="2025-10-05 07:14:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:26.42312141 +0000 UTC m=+1300.305747870" watchObservedRunningTime="2025-10-05 07:14:26.424228389 +0000 UTC m=+1300.306854849" Oct 05 07:14:26 crc kubenswrapper[4935]: I1005 07:14:26.797620 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59688ed9-ff1f-4c93-961c-247789e17af0" path="/var/lib/kubelet/pods/59688ed9-ff1f-4c93-961c-247789e17af0/volumes" Oct 05 07:14:27 crc kubenswrapper[4935]: I1005 07:14:27.416288 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9006ed82-292d-4af7-8851-fe468ea6f19a","Type":"ContainerStarted","Data":"779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17"} Oct 05 07:14:27 crc kubenswrapper[4935]: I1005 07:14:27.416576 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9006ed82-292d-4af7-8851-fe468ea6f19a","Type":"ContainerStarted","Data":"547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3"} Oct 05 07:14:27 crc kubenswrapper[4935]: I1005 07:14:27.445196 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.445175481 podStartE2EDuration="2.445175481s" podCreationTimestamp="2025-10-05 07:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:27.444929285 +0000 UTC m=+1301.327555765" watchObservedRunningTime="2025-10-05 07:14:27.445175481 +0000 UTC m=+1301.327801951" Oct 05 07:14:27 crc kubenswrapper[4935]: I1005 07:14:27.732492 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 05 07:14:29 crc kubenswrapper[4935]: I1005 07:14:29.694512 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 07:14:29 crc kubenswrapper[4935]: I1005 07:14:29.694833 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 07:14:29 crc kubenswrapper[4935]: I1005 07:14:29.843568 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 07:14:30 crc kubenswrapper[4935]: I1005 07:14:30.714107 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:14:30 crc kubenswrapper[4935]: I1005 07:14:30.714116 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:14:31 crc kubenswrapper[4935]: I1005 07:14:31.677365 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 07:14:34 crc kubenswrapper[4935]: I1005 07:14:34.842986 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 07:14:34 crc kubenswrapper[4935]: I1005 07:14:34.869797 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 07:14:34 crc kubenswrapper[4935]: I1005 07:14:34.958875 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:14:34 crc kubenswrapper[4935]: I1005 07:14:34.959096 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c0943221-d85f-4adb-a676-ffb7e7406eae" containerName="kube-state-metrics" containerID="cri-o://8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768" gracePeriod=30 Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.472144 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.516612 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c0943221-d85f-4adb-a676-ffb7e7406eae","Type":"ContainerDied","Data":"8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768"} Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.516685 4935 scope.go:117] "RemoveContainer" containerID="8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.516763 4935 generic.go:334] "Generic (PLEG): container finished" podID="c0943221-d85f-4adb-a676-ffb7e7406eae" containerID="8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768" exitCode=2 Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.516827 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.517436 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c0943221-d85f-4adb-a676-ffb7e7406eae","Type":"ContainerDied","Data":"0b29c4e44ab55818c0973083601100ca8f55ace514463250bc75a11b71f20891"} Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.530770 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b77dh\" (UniqueName: \"kubernetes.io/projected/c0943221-d85f-4adb-a676-ffb7e7406eae-kube-api-access-b77dh\") pod \"c0943221-d85f-4adb-a676-ffb7e7406eae\" (UID: \"c0943221-d85f-4adb-a676-ffb7e7406eae\") " Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.562139 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0943221-d85f-4adb-a676-ffb7e7406eae-kube-api-access-b77dh" (OuterVolumeSpecName: "kube-api-access-b77dh") pod "c0943221-d85f-4adb-a676-ffb7e7406eae" (UID: "c0943221-d85f-4adb-a676-ffb7e7406eae"). InnerVolumeSpecName "kube-api-access-b77dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.562324 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.562683 4935 scope.go:117] "RemoveContainer" containerID="8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768" Oct 05 07:14:35 crc kubenswrapper[4935]: E1005 07:14:35.563304 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768\": container with ID starting with 8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768 not found: ID does not exist" containerID="8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.563339 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768"} err="failed to get container status \"8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768\": rpc error: code = NotFound desc = could not find container \"8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768\": container with ID starting with 8b851beba350a72560aaba8e47520fbf366a6f15b7f8d00b41678d6f5a172768 not found: ID does not exist" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.642715 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b77dh\" (UniqueName: \"kubernetes.io/projected/c0943221-d85f-4adb-a676-ffb7e7406eae-kube-api-access-b77dh\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.870300 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.870342 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.891422 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.899030 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.916561 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:14:35 crc kubenswrapper[4935]: E1005 07:14:35.916984 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0943221-d85f-4adb-a676-ffb7e7406eae" containerName="kube-state-metrics" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.916999 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0943221-d85f-4adb-a676-ffb7e7406eae" containerName="kube-state-metrics" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.917170 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0943221-d85f-4adb-a676-ffb7e7406eae" containerName="kube-state-metrics" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.917760 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.920244 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.920851 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.926882 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.969241 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9jml\" (UniqueName: \"kubernetes.io/projected/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-api-access-d9jml\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.969286 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.969310 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:35 crc kubenswrapper[4935]: I1005 07:14:35.969376 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.070445 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9jml\" (UniqueName: \"kubernetes.io/projected/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-api-access-d9jml\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.070712 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.070839 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.070998 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.074476 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.074548 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.075373 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.096972 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9jml\" (UniqueName: \"kubernetes.io/projected/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-api-access-d9jml\") pod \"kube-state-metrics-0\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.236024 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.725381 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.727914 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="sg-core" containerID="cri-o://345c0040604dcbfb3be52da65117f237ffd6be39c38f28fb647f162dc6931d3f" gracePeriod=30 Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.728066 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="proxy-httpd" containerID="cri-o://696aa393f941516e4cfa1b451dd2f3e4440d4dee11149b8365d4707e731a8c95" gracePeriod=30 Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.728062 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="ceilometer-notification-agent" containerID="cri-o://ed56f6247798e566f4409b29c17aa572f76c05d02e825db11fb0cc73b7664132" gracePeriod=30 Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.728329 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="ceilometer-central-agent" containerID="cri-o://692ed475fb3453a3e6952fec05183b2dc6f37080fcd7dbbaf53b0d0a2e32e4ac" gracePeriod=30 Oct 05 07:14:36 crc kubenswrapper[4935]: W1005 07:14:36.766584 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc710b93b_c2fe_4a03_ba8c_f280a9d67da6.slice/crio-e3f24b7f7c6f483cb8122e57d065e46e001d54a174b2fb043d4bfb249336b0ee WatchSource:0}: Error finding container e3f24b7f7c6f483cb8122e57d065e46e001d54a174b2fb043d4bfb249336b0ee: Status 404 returned error can't find the container with id e3f24b7f7c6f483cb8122e57d065e46e001d54a174b2fb043d4bfb249336b0ee Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.773220 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.794476 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0943221-d85f-4adb-a676-ffb7e7406eae" path="/var/lib/kubelet/pods/c0943221-d85f-4adb-a676-ffb7e7406eae/volumes" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.911168 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 07:14:36 crc kubenswrapper[4935]: I1005 07:14:36.953136 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.540582 4935 generic.go:334] "Generic (PLEG): container finished" podID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerID="696aa393f941516e4cfa1b451dd2f3e4440d4dee11149b8365d4707e731a8c95" exitCode=0 Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.540974 4935 generic.go:334] "Generic (PLEG): container finished" podID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerID="345c0040604dcbfb3be52da65117f237ffd6be39c38f28fb647f162dc6931d3f" exitCode=2 Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.540986 4935 generic.go:334] "Generic (PLEG): container finished" podID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerID="692ed475fb3453a3e6952fec05183b2dc6f37080fcd7dbbaf53b0d0a2e32e4ac" exitCode=0 Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.540787 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerDied","Data":"696aa393f941516e4cfa1b451dd2f3e4440d4dee11149b8365d4707e731a8c95"} Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.541075 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerDied","Data":"345c0040604dcbfb3be52da65117f237ffd6be39c38f28fb647f162dc6931d3f"} Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.541093 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerDied","Data":"692ed475fb3453a3e6952fec05183b2dc6f37080fcd7dbbaf53b0d0a2e32e4ac"} Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.543871 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c710b93b-c2fe-4a03-ba8c-f280a9d67da6","Type":"ContainerStarted","Data":"da3d1180f3ad1963dd2711515fe79e75473abc8f27f2f472d500b4a81d01765f"} Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.543928 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c710b93b-c2fe-4a03-ba8c-f280a9d67da6","Type":"ContainerStarted","Data":"e3f24b7f7c6f483cb8122e57d065e46e001d54a174b2fb043d4bfb249336b0ee"} Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.544069 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 07:14:37 crc kubenswrapper[4935]: I1005 07:14:37.560222 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.131678382 podStartE2EDuration="2.56020668s" podCreationTimestamp="2025-10-05 07:14:35 +0000 UTC" firstStartedPulling="2025-10-05 07:14:36.770680681 +0000 UTC m=+1310.653307141" lastFinishedPulling="2025-10-05 07:14:37.199208989 +0000 UTC m=+1311.081835439" observedRunningTime="2025-10-05 07:14:37.557705404 +0000 UTC m=+1311.440331874" watchObservedRunningTime="2025-10-05 07:14:37.56020668 +0000 UTC m=+1311.442833140" Oct 05 07:14:39 crc kubenswrapper[4935]: I1005 07:14:39.699125 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 07:14:39 crc kubenswrapper[4935]: I1005 07:14:39.699475 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 07:14:39 crc kubenswrapper[4935]: I1005 07:14:39.705489 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 07:14:39 crc kubenswrapper[4935]: I1005 07:14:39.713883 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.574943 4935 generic.go:334] "Generic (PLEG): container finished" podID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerID="ed56f6247798e566f4409b29c17aa572f76c05d02e825db11fb0cc73b7664132" exitCode=0 Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.574937 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerDied","Data":"ed56f6247798e566f4409b29c17aa572f76c05d02e825db11fb0cc73b7664132"} Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.575423 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86d0c526-9ff7-4418-a237-92ab6a75b224","Type":"ContainerDied","Data":"7c044fe27b930c731a2e7a5d53f6c993308c009d5090fa1814f9075cfd63c1f5"} Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.575446 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c044fe27b930c731a2e7a5d53f6c993308c009d5090fa1814f9075cfd63c1f5" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.659492 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.774560 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-combined-ca-bundle\") pod \"86d0c526-9ff7-4418-a237-92ab6a75b224\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.774915 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-run-httpd\") pod \"86d0c526-9ff7-4418-a237-92ab6a75b224\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.774969 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-sg-core-conf-yaml\") pod \"86d0c526-9ff7-4418-a237-92ab6a75b224\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.775011 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-log-httpd\") pod \"86d0c526-9ff7-4418-a237-92ab6a75b224\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.775088 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-config-data\") pod \"86d0c526-9ff7-4418-a237-92ab6a75b224\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.775197 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwczs\" (UniqueName: \"kubernetes.io/projected/86d0c526-9ff7-4418-a237-92ab6a75b224-kube-api-access-hwczs\") pod \"86d0c526-9ff7-4418-a237-92ab6a75b224\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.775237 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-scripts\") pod \"86d0c526-9ff7-4418-a237-92ab6a75b224\" (UID: \"86d0c526-9ff7-4418-a237-92ab6a75b224\") " Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.775268 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "86d0c526-9ff7-4418-a237-92ab6a75b224" (UID: "86d0c526-9ff7-4418-a237-92ab6a75b224"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.775449 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "86d0c526-9ff7-4418-a237-92ab6a75b224" (UID: "86d0c526-9ff7-4418-a237-92ab6a75b224"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.775699 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.775715 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86d0c526-9ff7-4418-a237-92ab6a75b224-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.780245 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86d0c526-9ff7-4418-a237-92ab6a75b224-kube-api-access-hwczs" (OuterVolumeSpecName: "kube-api-access-hwczs") pod "86d0c526-9ff7-4418-a237-92ab6a75b224" (UID: "86d0c526-9ff7-4418-a237-92ab6a75b224"). InnerVolumeSpecName "kube-api-access-hwczs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.780752 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-scripts" (OuterVolumeSpecName: "scripts") pod "86d0c526-9ff7-4418-a237-92ab6a75b224" (UID: "86d0c526-9ff7-4418-a237-92ab6a75b224"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.813103 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "86d0c526-9ff7-4418-a237-92ab6a75b224" (UID: "86d0c526-9ff7-4418-a237-92ab6a75b224"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.871082 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-config-data" (OuterVolumeSpecName: "config-data") pod "86d0c526-9ff7-4418-a237-92ab6a75b224" (UID: "86d0c526-9ff7-4418-a237-92ab6a75b224"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.877467 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.877509 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.877526 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwczs\" (UniqueName: \"kubernetes.io/projected/86d0c526-9ff7-4418-a237-92ab6a75b224-kube-api-access-hwczs\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.877543 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.880082 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86d0c526-9ff7-4418-a237-92ab6a75b224" (UID: "86d0c526-9ff7-4418-a237-92ab6a75b224"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:40 crc kubenswrapper[4935]: I1005 07:14:40.979041 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0c526-9ff7-4418-a237-92ab6a75b224-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.585680 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.641048 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.651607 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.672344 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:41 crc kubenswrapper[4935]: E1005 07:14:41.675142 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="proxy-httpd" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.675229 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="proxy-httpd" Oct 05 07:14:41 crc kubenswrapper[4935]: E1005 07:14:41.675304 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="sg-core" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.675363 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="sg-core" Oct 05 07:14:41 crc kubenswrapper[4935]: E1005 07:14:41.675434 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="ceilometer-notification-agent" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.675494 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="ceilometer-notification-agent" Oct 05 07:14:41 crc kubenswrapper[4935]: E1005 07:14:41.675565 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="ceilometer-central-agent" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.675631 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="ceilometer-central-agent" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.675883 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="ceilometer-central-agent" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.675986 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="proxy-httpd" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.676088 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="ceilometer-notification-agent" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.676162 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" containerName="sg-core" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.678248 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.681003 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.681251 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.681383 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.686318 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.795723 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-config-data\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.795760 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-run-httpd\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.795778 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.795797 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-log-httpd\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.795842 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.795933 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.795974 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8tb5\" (UniqueName: \"kubernetes.io/projected/c1574459-bdf7-4b54-879f-ef0dd17cfff1-kube-api-access-b8tb5\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.796001 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-scripts\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.897463 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.897536 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8tb5\" (UniqueName: \"kubernetes.io/projected/c1574459-bdf7-4b54-879f-ef0dd17cfff1-kube-api-access-b8tb5\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.897567 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-scripts\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.897625 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-config-data\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.897641 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-run-httpd\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.897655 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.897671 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-log-httpd\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.897690 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.899282 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-log-httpd\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.899980 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-run-httpd\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.904093 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-config-data\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.904116 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.904939 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.911476 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-scripts\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.913481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:41 crc kubenswrapper[4935]: I1005 07:14:41.918748 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8tb5\" (UniqueName: \"kubernetes.io/projected/c1574459-bdf7-4b54-879f-ef0dd17cfff1-kube-api-access-b8tb5\") pod \"ceilometer-0\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " pod="openstack/ceilometer-0" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.006348 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.483464 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:42 crc kubenswrapper[4935]: W1005 07:14:42.486543 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1574459_bdf7_4b54_879f_ef0dd17cfff1.slice/crio-b075098bf77eb76a038040cd0a4fb55b68dc191ef493ee75b81c93a2dc4708f5 WatchSource:0}: Error finding container b075098bf77eb76a038040cd0a4fb55b68dc191ef493ee75b81c93a2dc4708f5: Status 404 returned error can't find the container with id b075098bf77eb76a038040cd0a4fb55b68dc191ef493ee75b81c93a2dc4708f5 Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.582518 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.593870 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerStarted","Data":"b075098bf77eb76a038040cd0a4fb55b68dc191ef493ee75b81c93a2dc4708f5"} Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.595520 4935 generic.go:334] "Generic (PLEG): container finished" podID="d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" containerID="13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e" exitCode=137 Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.595546 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e","Type":"ContainerDied","Data":"13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e"} Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.595561 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e","Type":"ContainerDied","Data":"3824cff43b8b7dffeb3caeefa0800f5c4d093949183cd04f9801f7534c259421"} Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.595577 4935 scope.go:117] "RemoveContainer" containerID="13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.595609 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.622020 4935 scope.go:117] "RemoveContainer" containerID="13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e" Oct 05 07:14:42 crc kubenswrapper[4935]: E1005 07:14:42.623169 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e\": container with ID starting with 13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e not found: ID does not exist" containerID="13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.623244 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e"} err="failed to get container status \"13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e\": rpc error: code = NotFound desc = could not find container \"13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e\": container with ID starting with 13aeaaf44e55dc9b3996bfa37d1e9ab2394e1e298ff75ea65fb631d8030eab6e not found: ID does not exist" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.716631 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-config-data\") pod \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.716809 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-combined-ca-bundle\") pod \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.716863 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbk5b\" (UniqueName: \"kubernetes.io/projected/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-kube-api-access-pbk5b\") pod \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\" (UID: \"d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e\") " Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.724911 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-kube-api-access-pbk5b" (OuterVolumeSpecName: "kube-api-access-pbk5b") pod "d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" (UID: "d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e"). InnerVolumeSpecName "kube-api-access-pbk5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.754844 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" (UID: "d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.756271 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-config-data" (OuterVolumeSpecName: "config-data") pod "d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" (UID: "d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.790218 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86d0c526-9ff7-4418-a237-92ab6a75b224" path="/var/lib/kubelet/pods/86d0c526-9ff7-4418-a237-92ab6a75b224/volumes" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.820297 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.821223 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.821312 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbk5b\" (UniqueName: \"kubernetes.io/projected/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e-kube-api-access-pbk5b\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.916853 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.926394 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.935341 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:42 crc kubenswrapper[4935]: E1005 07:14:42.935829 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.935855 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.936337 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.937019 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.941997 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.942219 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.942246 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 07:14:42 crc kubenswrapper[4935]: I1005 07:14:42.946800 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.024980 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7mb\" (UniqueName: \"kubernetes.io/projected/f950314f-27aa-4a44-b13a-4b4f3a7495ab-kube-api-access-zx7mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.025307 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.025688 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.025953 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.026153 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.128303 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.128712 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.128741 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.128836 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7mb\" (UniqueName: \"kubernetes.io/projected/f950314f-27aa-4a44-b13a-4b4f3a7495ab-kube-api-access-zx7mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.128865 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.134759 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.136215 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.136652 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.143607 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.145479 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7mb\" (UniqueName: \"kubernetes.io/projected/f950314f-27aa-4a44-b13a-4b4f3a7495ab-kube-api-access-zx7mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.254039 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.605282 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerStarted","Data":"3bd3805065ed5a284773fbd8d56f61ed68a99e16ba3f03746701e5ead0ba794b"} Oct 05 07:14:43 crc kubenswrapper[4935]: I1005 07:14:43.707947 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:14:43 crc kubenswrapper[4935]: W1005 07:14:43.721506 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf950314f_27aa_4a44_b13a_4b4f3a7495ab.slice/crio-b61e8ab4c7ba520380864b515362aca5c445949eca3b1e4d2053f94929164407 WatchSource:0}: Error finding container b61e8ab4c7ba520380864b515362aca5c445949eca3b1e4d2053f94929164407: Status 404 returned error can't find the container with id b61e8ab4c7ba520380864b515362aca5c445949eca3b1e4d2053f94929164407 Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.289476 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.289766 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.289818 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.290541 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b6d2f4c73d4ef3b7b40cc3341e20f2c6b3b4fe3d7affc8f4b8588270c63e170"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.290596 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://7b6d2f4c73d4ef3b7b40cc3341e20f2c6b3b4fe3d7affc8f4b8588270c63e170" gracePeriod=600 Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.616007 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="7b6d2f4c73d4ef3b7b40cc3341e20f2c6b3b4fe3d7affc8f4b8588270c63e170" exitCode=0 Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.616064 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"7b6d2f4c73d4ef3b7b40cc3341e20f2c6b3b4fe3d7affc8f4b8588270c63e170"} Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.616096 4935 scope.go:117] "RemoveContainer" containerID="be9c0311e869ecdb94c82f2824221de74d838dab99b4132c5a117c8e772b2a41" Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.617659 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f950314f-27aa-4a44-b13a-4b4f3a7495ab","Type":"ContainerStarted","Data":"1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd"} Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.617684 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f950314f-27aa-4a44-b13a-4b4f3a7495ab","Type":"ContainerStarted","Data":"b61e8ab4c7ba520380864b515362aca5c445949eca3b1e4d2053f94929164407"} Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.620405 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerStarted","Data":"b9199efbd18b8014951e355465e8d05c3b1153672e8bb9f72930804c0201e367"} Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.659924 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.659903111 podStartE2EDuration="2.659903111s" podCreationTimestamp="2025-10-05 07:14:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:44.64060206 +0000 UTC m=+1318.523228530" watchObservedRunningTime="2025-10-05 07:14:44.659903111 +0000 UTC m=+1318.542529571" Oct 05 07:14:44 crc kubenswrapper[4935]: I1005 07:14:44.788122 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e" path="/var/lib/kubelet/pods/d6caa1e4-cc0f-4abd-ada3-5f1e4e96a51e/volumes" Oct 05 07:14:45 crc kubenswrapper[4935]: I1005 07:14:45.631543 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072"} Oct 05 07:14:45 crc kubenswrapper[4935]: I1005 07:14:45.636031 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerStarted","Data":"6ec7f40ae44ce5a7a27563f6b28c6a59c3d126a6e4af11f10ced3dc5e4b2adc6"} Oct 05 07:14:45 crc kubenswrapper[4935]: I1005 07:14:45.891618 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 07:14:45 crc kubenswrapper[4935]: I1005 07:14:45.891730 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 07:14:45 crc kubenswrapper[4935]: I1005 07:14:45.893018 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 07:14:45 crc kubenswrapper[4935]: I1005 07:14:45.893056 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 07:14:45 crc kubenswrapper[4935]: I1005 07:14:45.909461 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 07:14:45 crc kubenswrapper[4935]: I1005 07:14:45.910194 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.132646 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-47z4r"] Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.135194 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.151719 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-47z4r"] Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.225556 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-swift-storage-0\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.225640 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-config\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.225665 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-svc\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.225707 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.225737 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.225820 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4fxg\" (UniqueName: \"kubernetes.io/projected/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-kube-api-access-b4fxg\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.247587 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.327760 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-swift-storage-0\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.327845 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-config\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.327869 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-svc\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.328642 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-swift-storage-0\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.328775 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-config\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.329328 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-svc\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.329376 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.329419 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.329514 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4fxg\" (UniqueName: \"kubernetes.io/projected/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-kube-api-access-b4fxg\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.331445 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.331469 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.356763 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4fxg\" (UniqueName: \"kubernetes.io/projected/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-kube-api-access-b4fxg\") pod \"dnsmasq-dns-6c5b8b845f-47z4r\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.463253 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.657548 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerStarted","Data":"dc51f8c2d9057313e4971d14237dd2010027334ea19793e2c0191ed7fc55fb03"} Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.687950 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.350399016 podStartE2EDuration="5.687755444s" podCreationTimestamp="2025-10-05 07:14:41 +0000 UTC" firstStartedPulling="2025-10-05 07:14:42.498115926 +0000 UTC m=+1316.380742386" lastFinishedPulling="2025-10-05 07:14:45.835472354 +0000 UTC m=+1319.718098814" observedRunningTime="2025-10-05 07:14:46.675975172 +0000 UTC m=+1320.558601632" watchObservedRunningTime="2025-10-05 07:14:46.687755444 +0000 UTC m=+1320.570381904" Oct 05 07:14:46 crc kubenswrapper[4935]: W1005 07:14:46.942996 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bf67c05_04fb_414a_aa6e_8cdc3831cb94.slice/crio-114eb74ef5792ce6ed05052372361ebe58fe546bc2e8963abcc523b948677238 WatchSource:0}: Error finding container 114eb74ef5792ce6ed05052372361ebe58fe546bc2e8963abcc523b948677238: Status 404 returned error can't find the container with id 114eb74ef5792ce6ed05052372361ebe58fe546bc2e8963abcc523b948677238 Oct 05 07:14:46 crc kubenswrapper[4935]: I1005 07:14:46.962399 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-47z4r"] Oct 05 07:14:47 crc kubenswrapper[4935]: I1005 07:14:47.605004 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:47 crc kubenswrapper[4935]: I1005 07:14:47.667127 4935 generic.go:334] "Generic (PLEG): container finished" podID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerID="90e177c52c5d7180cd73793a31e9ceb0eeac1791a8bf7d225dc5c4a85800b727" exitCode=0 Oct 05 07:14:47 crc kubenswrapper[4935]: I1005 07:14:47.667222 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" event={"ID":"8bf67c05-04fb-414a-aa6e-8cdc3831cb94","Type":"ContainerDied","Data":"90e177c52c5d7180cd73793a31e9ceb0eeac1791a8bf7d225dc5c4a85800b727"} Oct 05 07:14:47 crc kubenswrapper[4935]: I1005 07:14:47.667262 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" event={"ID":"8bf67c05-04fb-414a-aa6e-8cdc3831cb94","Type":"ContainerStarted","Data":"114eb74ef5792ce6ed05052372361ebe58fe546bc2e8963abcc523b948677238"} Oct 05 07:14:47 crc kubenswrapper[4935]: I1005 07:14:47.667448 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.255088 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.675520 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="ceilometer-central-agent" containerID="cri-o://3bd3805065ed5a284773fbd8d56f61ed68a99e16ba3f03746701e5ead0ba794b" gracePeriod=30 Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.676582 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" event={"ID":"8bf67c05-04fb-414a-aa6e-8cdc3831cb94","Type":"ContainerStarted","Data":"5c401639335e92358e96e3ee9d7c9cbcdd83e9b6aa688f56a6c57891805ffe6d"} Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.676613 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.676838 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="proxy-httpd" containerID="cri-o://dc51f8c2d9057313e4971d14237dd2010027334ea19793e2c0191ed7fc55fb03" gracePeriod=30 Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.676884 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="sg-core" containerID="cri-o://6ec7f40ae44ce5a7a27563f6b28c6a59c3d126a6e4af11f10ced3dc5e4b2adc6" gracePeriod=30 Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.676933 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="ceilometer-notification-agent" containerID="cri-o://b9199efbd18b8014951e355465e8d05c3b1153672e8bb9f72930804c0201e367" gracePeriod=30 Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.707098 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" podStartSLOduration=2.70708015 podStartE2EDuration="2.70708015s" podCreationTimestamp="2025-10-05 07:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:48.696179381 +0000 UTC m=+1322.578805841" watchObservedRunningTime="2025-10-05 07:14:48.70708015 +0000 UTC m=+1322.589706610" Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.957253 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.957482 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-log" containerID="cri-o://547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3" gracePeriod=30 Oct 05 07:14:48 crc kubenswrapper[4935]: I1005 07:14:48.957539 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-api" containerID="cri-o://779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17" gracePeriod=30 Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.685508 4935 generic.go:334] "Generic (PLEG): container finished" podID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerID="547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3" exitCode=143 Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.685675 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9006ed82-292d-4af7-8851-fe468ea6f19a","Type":"ContainerDied","Data":"547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3"} Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.689248 4935 generic.go:334] "Generic (PLEG): container finished" podID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerID="dc51f8c2d9057313e4971d14237dd2010027334ea19793e2c0191ed7fc55fb03" exitCode=0 Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.689273 4935 generic.go:334] "Generic (PLEG): container finished" podID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerID="6ec7f40ae44ce5a7a27563f6b28c6a59c3d126a6e4af11f10ced3dc5e4b2adc6" exitCode=2 Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.689280 4935 generic.go:334] "Generic (PLEG): container finished" podID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerID="b9199efbd18b8014951e355465e8d05c3b1153672e8bb9f72930804c0201e367" exitCode=0 Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.689286 4935 generic.go:334] "Generic (PLEG): container finished" podID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerID="3bd3805065ed5a284773fbd8d56f61ed68a99e16ba3f03746701e5ead0ba794b" exitCode=0 Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.689973 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerDied","Data":"dc51f8c2d9057313e4971d14237dd2010027334ea19793e2c0191ed7fc55fb03"} Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.690002 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerDied","Data":"6ec7f40ae44ce5a7a27563f6b28c6a59c3d126a6e4af11f10ced3dc5e4b2adc6"} Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.690013 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerDied","Data":"b9199efbd18b8014951e355465e8d05c3b1153672e8bb9f72930804c0201e367"} Oct 05 07:14:49 crc kubenswrapper[4935]: I1005 07:14:49.690021 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerDied","Data":"3bd3805065ed5a284773fbd8d56f61ed68a99e16ba3f03746701e5ead0ba794b"} Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.471655 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.624707 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-ceilometer-tls-certs\") pod \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.625098 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-combined-ca-bundle\") pod \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.625243 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-log-httpd\") pod \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.625343 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-config-data\") pod \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.625442 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8tb5\" (UniqueName: \"kubernetes.io/projected/c1574459-bdf7-4b54-879f-ef0dd17cfff1-kube-api-access-b8tb5\") pod \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.625563 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-scripts\") pod \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.625664 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-sg-core-conf-yaml\") pod \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.625765 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-run-httpd\") pod \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\" (UID: \"c1574459-bdf7-4b54-879f-ef0dd17cfff1\") " Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.625757 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c1574459-bdf7-4b54-879f-ef0dd17cfff1" (UID: "c1574459-bdf7-4b54-879f-ef0dd17cfff1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.626046 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c1574459-bdf7-4b54-879f-ef0dd17cfff1" (UID: "c1574459-bdf7-4b54-879f-ef0dd17cfff1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.626367 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.626395 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1574459-bdf7-4b54-879f-ef0dd17cfff1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.630197 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1574459-bdf7-4b54-879f-ef0dd17cfff1-kube-api-access-b8tb5" (OuterVolumeSpecName: "kube-api-access-b8tb5") pod "c1574459-bdf7-4b54-879f-ef0dd17cfff1" (UID: "c1574459-bdf7-4b54-879f-ef0dd17cfff1"). InnerVolumeSpecName "kube-api-access-b8tb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.630707 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-scripts" (OuterVolumeSpecName: "scripts") pod "c1574459-bdf7-4b54-879f-ef0dd17cfff1" (UID: "c1574459-bdf7-4b54-879f-ef0dd17cfff1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.650439 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c1574459-bdf7-4b54-879f-ef0dd17cfff1" (UID: "c1574459-bdf7-4b54-879f-ef0dd17cfff1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.683239 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c1574459-bdf7-4b54-879f-ef0dd17cfff1" (UID: "c1574459-bdf7-4b54-879f-ef0dd17cfff1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.721076 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1574459-bdf7-4b54-879f-ef0dd17cfff1","Type":"ContainerDied","Data":"b075098bf77eb76a038040cd0a4fb55b68dc191ef493ee75b81c93a2dc4708f5"} Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.721182 4935 scope.go:117] "RemoveContainer" containerID="dc51f8c2d9057313e4971d14237dd2010027334ea19793e2c0191ed7fc55fb03" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.721359 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.727483 4935 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.727510 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8tb5\" (UniqueName: \"kubernetes.io/projected/c1574459-bdf7-4b54-879f-ef0dd17cfff1-kube-api-access-b8tb5\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.727522 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.727531 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.752275 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1574459-bdf7-4b54-879f-ef0dd17cfff1" (UID: "c1574459-bdf7-4b54-879f-ef0dd17cfff1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.767007 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-config-data" (OuterVolumeSpecName: "config-data") pod "c1574459-bdf7-4b54-879f-ef0dd17cfff1" (UID: "c1574459-bdf7-4b54-879f-ef0dd17cfff1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.815346 4935 scope.go:117] "RemoveContainer" containerID="6ec7f40ae44ce5a7a27563f6b28c6a59c3d126a6e4af11f10ced3dc5e4b2adc6" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.828678 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.828968 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1574459-bdf7-4b54-879f-ef0dd17cfff1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.839125 4935 scope.go:117] "RemoveContainer" containerID="b9199efbd18b8014951e355465e8d05c3b1153672e8bb9f72930804c0201e367" Oct 05 07:14:50 crc kubenswrapper[4935]: I1005 07:14:50.857841 4935 scope.go:117] "RemoveContainer" containerID="3bd3805065ed5a284773fbd8d56f61ed68a99e16ba3f03746701e5ead0ba794b" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.055973 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.077271 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.090802 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:51 crc kubenswrapper[4935]: E1005 07:14:51.091238 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="ceilometer-notification-agent" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.091250 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="ceilometer-notification-agent" Oct 05 07:14:51 crc kubenswrapper[4935]: E1005 07:14:51.091267 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="proxy-httpd" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.091273 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="proxy-httpd" Oct 05 07:14:51 crc kubenswrapper[4935]: E1005 07:14:51.091287 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="ceilometer-central-agent" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.091292 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="ceilometer-central-agent" Oct 05 07:14:51 crc kubenswrapper[4935]: E1005 07:14:51.091306 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="sg-core" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.091312 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="sg-core" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.091482 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="sg-core" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.091497 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="proxy-httpd" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.091506 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="ceilometer-notification-agent" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.091513 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" containerName="ceilometer-central-agent" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.093236 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.097741 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.097915 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.098025 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.098425 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.236367 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-config-data\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.236576 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-log-httpd\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.236956 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-run-httpd\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.237060 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.237413 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.237478 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.237588 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-scripts\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.237622 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sj49\" (UniqueName: \"kubernetes.io/projected/861347ca-4aba-4574-b5f8-fd18fb13c27e-kube-api-access-4sj49\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.252335 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:51 crc kubenswrapper[4935]: E1005 07:14:51.253663 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-4sj49 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="861347ca-4aba-4574-b5f8-fd18fb13c27e" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.338947 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-scripts\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.338983 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sj49\" (UniqueName: \"kubernetes.io/projected/861347ca-4aba-4574-b5f8-fd18fb13c27e-kube-api-access-4sj49\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.339032 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-config-data\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.339068 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-log-httpd\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.339116 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-run-httpd\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.339132 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.339205 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.339220 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.339826 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-run-httpd\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.340026 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-log-httpd\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.342464 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.343385 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-scripts\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.344204 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-config-data\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.346240 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.346843 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.358189 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sj49\" (UniqueName: \"kubernetes.io/projected/861347ca-4aba-4574-b5f8-fd18fb13c27e-kube-api-access-4sj49\") pod \"ceilometer-0\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.731771 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.753583 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.848817 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-scripts\") pod \"861347ca-4aba-4574-b5f8-fd18fb13c27e\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.848857 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-ceilometer-tls-certs\") pod \"861347ca-4aba-4574-b5f8-fd18fb13c27e\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.848930 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-combined-ca-bundle\") pod \"861347ca-4aba-4574-b5f8-fd18fb13c27e\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.848994 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-config-data\") pod \"861347ca-4aba-4574-b5f8-fd18fb13c27e\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.849095 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-run-httpd\") pod \"861347ca-4aba-4574-b5f8-fd18fb13c27e\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.849156 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-log-httpd\") pod \"861347ca-4aba-4574-b5f8-fd18fb13c27e\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.849202 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sj49\" (UniqueName: \"kubernetes.io/projected/861347ca-4aba-4574-b5f8-fd18fb13c27e-kube-api-access-4sj49\") pod \"861347ca-4aba-4574-b5f8-fd18fb13c27e\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.849239 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-sg-core-conf-yaml\") pod \"861347ca-4aba-4574-b5f8-fd18fb13c27e\" (UID: \"861347ca-4aba-4574-b5f8-fd18fb13c27e\") " Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.849707 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "861347ca-4aba-4574-b5f8-fd18fb13c27e" (UID: "861347ca-4aba-4574-b5f8-fd18fb13c27e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.849787 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "861347ca-4aba-4574-b5f8-fd18fb13c27e" (UID: "861347ca-4aba-4574-b5f8-fd18fb13c27e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.853010 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "861347ca-4aba-4574-b5f8-fd18fb13c27e" (UID: "861347ca-4aba-4574-b5f8-fd18fb13c27e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.853725 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "861347ca-4aba-4574-b5f8-fd18fb13c27e" (UID: "861347ca-4aba-4574-b5f8-fd18fb13c27e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.854289 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-config-data" (OuterVolumeSpecName: "config-data") pod "861347ca-4aba-4574-b5f8-fd18fb13c27e" (UID: "861347ca-4aba-4574-b5f8-fd18fb13c27e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.854332 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "861347ca-4aba-4574-b5f8-fd18fb13c27e" (UID: "861347ca-4aba-4574-b5f8-fd18fb13c27e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.854361 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-scripts" (OuterVolumeSpecName: "scripts") pod "861347ca-4aba-4574-b5f8-fd18fb13c27e" (UID: "861347ca-4aba-4574-b5f8-fd18fb13c27e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.854304 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/861347ca-4aba-4574-b5f8-fd18fb13c27e-kube-api-access-4sj49" (OuterVolumeSpecName: "kube-api-access-4sj49") pod "861347ca-4aba-4574-b5f8-fd18fb13c27e" (UID: "861347ca-4aba-4574-b5f8-fd18fb13c27e"). InnerVolumeSpecName "kube-api-access-4sj49". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.952080 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.952401 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.952683 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/861347ca-4aba-4574-b5f8-fd18fb13c27e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.952820 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sj49\" (UniqueName: \"kubernetes.io/projected/861347ca-4aba-4574-b5f8-fd18fb13c27e-kube-api-access-4sj49\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.952982 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.953152 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.953292 4935 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:51 crc kubenswrapper[4935]: I1005 07:14:51.953421 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861347ca-4aba-4574-b5f8-fd18fb13c27e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.680944 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.746838 4935 generic.go:334] "Generic (PLEG): container finished" podID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerID="779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17" exitCode=0 Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.746928 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.747075 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9006ed82-292d-4af7-8851-fe468ea6f19a","Type":"ContainerDied","Data":"779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17"} Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.747131 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9006ed82-292d-4af7-8851-fe468ea6f19a","Type":"ContainerDied","Data":"877a2a14723c5e8aefa38d878377ad7723232f520dfcbca833472a4560d39fe7"} Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.747158 4935 scope.go:117] "RemoveContainer" containerID="779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.748152 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.806209 4935 scope.go:117] "RemoveContainer" containerID="547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.810287 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1574459-bdf7-4b54-879f-ef0dd17cfff1" path="/var/lib/kubelet/pods/c1574459-bdf7-4b54-879f-ef0dd17cfff1/volumes" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.830175 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.830222 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.844976 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:52 crc kubenswrapper[4935]: E1005 07:14:52.845408 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-log" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.845425 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-log" Oct 05 07:14:52 crc kubenswrapper[4935]: E1005 07:14:52.845444 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-api" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.845453 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-api" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.845863 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-api" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.845924 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" containerName="nova-api-log" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.872667 4935 scope.go:117] "RemoveContainer" containerID="779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17" Oct 05 07:14:52 crc kubenswrapper[4935]: E1005 07:14:52.873159 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17\": container with ID starting with 779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17 not found: ID does not exist" containerID="779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.873332 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17"} err="failed to get container status \"779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17\": rpc error: code = NotFound desc = could not find container \"779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17\": container with ID starting with 779d824699299083686b0c8f41e89be6692f446dd627ef0d2610d5ab5ce5fe17 not found: ID does not exist" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.873577 4935 scope.go:117] "RemoveContainer" containerID="547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3" Oct 05 07:14:52 crc kubenswrapper[4935]: E1005 07:14:52.874369 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3\": container with ID starting with 547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3 not found: ID does not exist" containerID="547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.874413 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3"} err="failed to get container status \"547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3\": rpc error: code = NotFound desc = could not find container \"547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3\": container with ID starting with 547253e3f2c4d4d04f7b081cff53300ec7215fc7c717ce0b65b6c4b60753bef3 not found: ID does not exist" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.874660 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9006ed82-292d-4af7-8851-fe468ea6f19a-logs\") pod \"9006ed82-292d-4af7-8851-fe468ea6f19a\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.874718 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-combined-ca-bundle\") pod \"9006ed82-292d-4af7-8851-fe468ea6f19a\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.874881 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-config-data\") pod \"9006ed82-292d-4af7-8851-fe468ea6f19a\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.874955 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cklk\" (UniqueName: \"kubernetes.io/projected/9006ed82-292d-4af7-8851-fe468ea6f19a-kube-api-access-4cklk\") pod \"9006ed82-292d-4af7-8851-fe468ea6f19a\" (UID: \"9006ed82-292d-4af7-8851-fe468ea6f19a\") " Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.884010 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9006ed82-292d-4af7-8851-fe468ea6f19a-logs" (OuterVolumeSpecName: "logs") pod "9006ed82-292d-4af7-8851-fe468ea6f19a" (UID: "9006ed82-292d-4af7-8851-fe468ea6f19a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.887722 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.887859 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.888147 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9006ed82-292d-4af7-8851-fe468ea6f19a-kube-api-access-4cklk" (OuterVolumeSpecName: "kube-api-access-4cklk") pod "9006ed82-292d-4af7-8851-fe468ea6f19a" (UID: "9006ed82-292d-4af7-8851-fe468ea6f19a"). InnerVolumeSpecName "kube-api-access-4cklk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.889840 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.892583 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.892582 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.907695 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9006ed82-292d-4af7-8851-fe468ea6f19a" (UID: "9006ed82-292d-4af7-8851-fe468ea6f19a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.919958 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-config-data" (OuterVolumeSpecName: "config-data") pod "9006ed82-292d-4af7-8851-fe468ea6f19a" (UID: "9006ed82-292d-4af7-8851-fe468ea6f19a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.977161 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-log-httpd\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.977454 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.977710 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-config-data\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.977910 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-scripts\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.977991 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.978038 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-run-httpd\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.978059 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.978084 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs7jr\" (UniqueName: \"kubernetes.io/projected/aba39a0b-f275-4075-9419-a354a3c5fa5e-kube-api-access-rs7jr\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.978277 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.978325 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9006ed82-292d-4af7-8851-fe468ea6f19a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.978346 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cklk\" (UniqueName: \"kubernetes.io/projected/9006ed82-292d-4af7-8851-fe468ea6f19a-kube-api-access-4cklk\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:52 crc kubenswrapper[4935]: I1005 07:14:52.978367 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9006ed82-292d-4af7-8851-fe468ea6f19a-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.076221 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.079881 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-scripts\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.079987 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.080033 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-run-httpd\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.080068 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.080106 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs7jr\" (UniqueName: \"kubernetes.io/projected/aba39a0b-f275-4075-9419-a354a3c5fa5e-kube-api-access-rs7jr\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.080193 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-log-httpd\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.080264 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.080462 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-config-data\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.080501 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-run-httpd\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.080871 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-log-httpd\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.084487 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.084628 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-config-data\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.084494 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-scripts\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.085253 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.088738 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.090960 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.105456 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs7jr\" (UniqueName: \"kubernetes.io/projected/aba39a0b-f275-4075-9419-a354a3c5fa5e-kube-api-access-rs7jr\") pod \"ceilometer-0\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.112982 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.114769 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.116437 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.116481 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.116715 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.122678 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.255084 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.278336 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.278648 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.288495 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gjb2\" (UniqueName: \"kubernetes.io/projected/78323dc0-73ec-47d4-bdf0-639a3f308db1-kube-api-access-9gjb2\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.288573 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-config-data\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.288665 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.288685 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78323dc0-73ec-47d4-bdf0-639a3f308db1-logs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.288761 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-public-tls-certs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.288796 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.390257 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gjb2\" (UniqueName: \"kubernetes.io/projected/78323dc0-73ec-47d4-bdf0-639a3f308db1-kube-api-access-9gjb2\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.390316 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-config-data\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.390341 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.390361 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78323dc0-73ec-47d4-bdf0-639a3f308db1-logs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.390443 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-public-tls-certs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.390462 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.395701 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-config-data\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.397005 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.397103 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78323dc0-73ec-47d4-bdf0-639a3f308db1-logs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.397748 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.414563 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-public-tls-certs\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.418644 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gjb2\" (UniqueName: \"kubernetes.io/projected/78323dc0-73ec-47d4-bdf0-639a3f308db1-kube-api-access-9gjb2\") pod \"nova-api-0\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.461626 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.729553 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.758042 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerStarted","Data":"cfed2e35aa6cff13d197095c51fbf6bf35c8a059124e7708c37f3b502f0eed87"} Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.780785 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.893737 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:14:53 crc kubenswrapper[4935]: W1005 07:14:53.895186 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78323dc0_73ec_47d4_bdf0_639a3f308db1.slice/crio-f1f9b8b1fd64c05287d4d21a006676608b0375909ca8454e32a4b8de48a4e9cd WatchSource:0}: Error finding container f1f9b8b1fd64c05287d4d21a006676608b0375909ca8454e32a4b8de48a4e9cd: Status 404 returned error can't find the container with id f1f9b8b1fd64c05287d4d21a006676608b0375909ca8454e32a4b8de48a4e9cd Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.973005 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-bm525"] Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.974448 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.976630 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 05 07:14:53 crc kubenswrapper[4935]: I1005 07:14:53.977445 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.002739 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bm525"] Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.104028 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw4zh\" (UniqueName: \"kubernetes.io/projected/fbda9936-7e2e-49ec-ad44-29f6451a266b-kube-api-access-fw4zh\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.104103 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.104146 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-config-data\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.104218 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-scripts\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.205839 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw4zh\" (UniqueName: \"kubernetes.io/projected/fbda9936-7e2e-49ec-ad44-29f6451a266b-kube-api-access-fw4zh\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.206244 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.206281 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-config-data\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.206327 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-scripts\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.214488 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.215160 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-config-data\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.217004 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-scripts\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.221745 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw4zh\" (UniqueName: \"kubernetes.io/projected/fbda9936-7e2e-49ec-ad44-29f6451a266b-kube-api-access-fw4zh\") pod \"nova-cell1-cell-mapping-bm525\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.310985 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.774155 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78323dc0-73ec-47d4-bdf0-639a3f308db1","Type":"ContainerStarted","Data":"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2"} Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.812257 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.812236387 podStartE2EDuration="1.812236387s" podCreationTimestamp="2025-10-05 07:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:54.795639758 +0000 UTC m=+1328.678266228" watchObservedRunningTime="2025-10-05 07:14:54.812236387 +0000 UTC m=+1328.694862857" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.815409 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="861347ca-4aba-4574-b5f8-fd18fb13c27e" path="/var/lib/kubelet/pods/861347ca-4aba-4574-b5f8-fd18fb13c27e/volumes" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.815812 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9006ed82-292d-4af7-8851-fe468ea6f19a" path="/var/lib/kubelet/pods/9006ed82-292d-4af7-8851-fe468ea6f19a/volumes" Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.816542 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bm525"] Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.816568 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78323dc0-73ec-47d4-bdf0-639a3f308db1","Type":"ContainerStarted","Data":"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da"} Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.817050 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78323dc0-73ec-47d4-bdf0-639a3f308db1","Type":"ContainerStarted","Data":"f1f9b8b1fd64c05287d4d21a006676608b0375909ca8454e32a4b8de48a4e9cd"} Oct 05 07:14:54 crc kubenswrapper[4935]: I1005 07:14:54.817078 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerStarted","Data":"1a7bd0c309de5a9b869370ed377d8a46c8a2f0c566fa6724d9ed2c87fbfb0e4a"} Oct 05 07:14:55 crc kubenswrapper[4935]: I1005 07:14:55.793384 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerStarted","Data":"48ae75168e8a69d6f8a97da4dacb4141e0980a9ae4a1946601d92425b5f4f413"} Oct 05 07:14:55 crc kubenswrapper[4935]: I1005 07:14:55.793971 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerStarted","Data":"37c00f2d3e9897fa787e5b2b4870c3b4e33ad0c59b0775763a7aeb443b6567b7"} Oct 05 07:14:55 crc kubenswrapper[4935]: I1005 07:14:55.797308 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bm525" event={"ID":"fbda9936-7e2e-49ec-ad44-29f6451a266b","Type":"ContainerStarted","Data":"98baaaf4f3d464e0f5da3244138847077c9066de79336a50452f980e1c7db9d4"} Oct 05 07:14:55 crc kubenswrapper[4935]: I1005 07:14:55.797384 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bm525" event={"ID":"fbda9936-7e2e-49ec-ad44-29f6451a266b","Type":"ContainerStarted","Data":"785cb990cbcda9b1211e614b1551e11272ec279717ac97e3c3ffb24982e54750"} Oct 05 07:14:55 crc kubenswrapper[4935]: I1005 07:14:55.817937 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-bm525" podStartSLOduration=2.817906325 podStartE2EDuration="2.817906325s" podCreationTimestamp="2025-10-05 07:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:14:55.811206658 +0000 UTC m=+1329.693833118" watchObservedRunningTime="2025-10-05 07:14:55.817906325 +0000 UTC m=+1329.700532785" Oct 05 07:14:56 crc kubenswrapper[4935]: I1005 07:14:56.465334 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:14:56 crc kubenswrapper[4935]: I1005 07:14:56.530648 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-5qg9s"] Oct 05 07:14:56 crc kubenswrapper[4935]: I1005 07:14:56.530931 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" podUID="a02edeb1-77e8-4884-a922-3dc7d608b587" containerName="dnsmasq-dns" containerID="cri-o://9dd6b516cda99fd95d5c309707235a54553d259bf1fba2f5808a58484cacac39" gracePeriod=10 Oct 05 07:14:56 crc kubenswrapper[4935]: I1005 07:14:56.832935 4935 generic.go:334] "Generic (PLEG): container finished" podID="a02edeb1-77e8-4884-a922-3dc7d608b587" containerID="9dd6b516cda99fd95d5c309707235a54553d259bf1fba2f5808a58484cacac39" exitCode=0 Oct 05 07:14:56 crc kubenswrapper[4935]: I1005 07:14:56.833935 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" event={"ID":"a02edeb1-77e8-4884-a922-3dc7d608b587","Type":"ContainerDied","Data":"9dd6b516cda99fd95d5c309707235a54553d259bf1fba2f5808a58484cacac39"} Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.118369 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.277360 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-nb\") pod \"a02edeb1-77e8-4884-a922-3dc7d608b587\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.277831 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-sb\") pod \"a02edeb1-77e8-4884-a922-3dc7d608b587\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.277855 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-config\") pod \"a02edeb1-77e8-4884-a922-3dc7d608b587\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.277982 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n96vp\" (UniqueName: \"kubernetes.io/projected/a02edeb1-77e8-4884-a922-3dc7d608b587-kube-api-access-n96vp\") pod \"a02edeb1-77e8-4884-a922-3dc7d608b587\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.278030 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-svc\") pod \"a02edeb1-77e8-4884-a922-3dc7d608b587\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.278133 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-swift-storage-0\") pod \"a02edeb1-77e8-4884-a922-3dc7d608b587\" (UID: \"a02edeb1-77e8-4884-a922-3dc7d608b587\") " Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.294037 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a02edeb1-77e8-4884-a922-3dc7d608b587-kube-api-access-n96vp" (OuterVolumeSpecName: "kube-api-access-n96vp") pod "a02edeb1-77e8-4884-a922-3dc7d608b587" (UID: "a02edeb1-77e8-4884-a922-3dc7d608b587"). InnerVolumeSpecName "kube-api-access-n96vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.340372 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a02edeb1-77e8-4884-a922-3dc7d608b587" (UID: "a02edeb1-77e8-4884-a922-3dc7d608b587"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.348265 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-config" (OuterVolumeSpecName: "config") pod "a02edeb1-77e8-4884-a922-3dc7d608b587" (UID: "a02edeb1-77e8-4884-a922-3dc7d608b587"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.348628 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a02edeb1-77e8-4884-a922-3dc7d608b587" (UID: "a02edeb1-77e8-4884-a922-3dc7d608b587"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.354207 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a02edeb1-77e8-4884-a922-3dc7d608b587" (UID: "a02edeb1-77e8-4884-a922-3dc7d608b587"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.379640 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.379672 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.379695 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n96vp\" (UniqueName: \"kubernetes.io/projected/a02edeb1-77e8-4884-a922-3dc7d608b587-kube-api-access-n96vp\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.379711 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.379722 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.381716 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a02edeb1-77e8-4884-a922-3dc7d608b587" (UID: "a02edeb1-77e8-4884-a922-3dc7d608b587"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.482061 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a02edeb1-77e8-4884-a922-3dc7d608b587-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.844264 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" event={"ID":"a02edeb1-77e8-4884-a922-3dc7d608b587","Type":"ContainerDied","Data":"bebcce117c4b9d97cca2915d750877ff31291775637b77a44b0d41b150a80cb9"} Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.844300 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6d46f6cf-5qg9s" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.844350 4935 scope.go:117] "RemoveContainer" containerID="9dd6b516cda99fd95d5c309707235a54553d259bf1fba2f5808a58484cacac39" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.846990 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerStarted","Data":"fc4511ccd783fd2b8c6cb4dce5c4f3cc88512854346a9a9f11bfe9081fb534e4"} Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.847267 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.899750 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.765910682 podStartE2EDuration="5.899731825s" podCreationTimestamp="2025-10-05 07:14:52 +0000 UTC" firstStartedPulling="2025-10-05 07:14:53.733587419 +0000 UTC m=+1327.616213889" lastFinishedPulling="2025-10-05 07:14:56.867408572 +0000 UTC m=+1330.750035032" observedRunningTime="2025-10-05 07:14:57.87950015 +0000 UTC m=+1331.762126620" watchObservedRunningTime="2025-10-05 07:14:57.899731825 +0000 UTC m=+1331.782358285" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.900728 4935 scope.go:117] "RemoveContainer" containerID="95737c65c8ee191ece2d2590b4f6194158f5c08fb6949f2e857aa29e0037e4c3" Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.921011 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-5qg9s"] Oct 05 07:14:57 crc kubenswrapper[4935]: I1005 07:14:57.929115 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-5qg9s"] Oct 05 07:14:58 crc kubenswrapper[4935]: I1005 07:14:58.791727 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a02edeb1-77e8-4884-a922-3dc7d608b587" path="/var/lib/kubelet/pods/a02edeb1-77e8-4884-a922-3dc7d608b587/volumes" Oct 05 07:14:59 crc kubenswrapper[4935]: I1005 07:14:59.871619 4935 generic.go:334] "Generic (PLEG): container finished" podID="fbda9936-7e2e-49ec-ad44-29f6451a266b" containerID="98baaaf4f3d464e0f5da3244138847077c9066de79336a50452f980e1c7db9d4" exitCode=0 Oct 05 07:14:59 crc kubenswrapper[4935]: I1005 07:14:59.871715 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bm525" event={"ID":"fbda9936-7e2e-49ec-ad44-29f6451a266b","Type":"ContainerDied","Data":"98baaaf4f3d464e0f5da3244138847077c9066de79336a50452f980e1c7db9d4"} Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.153409 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz"] Oct 05 07:15:00 crc kubenswrapper[4935]: E1005 07:15:00.154027 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a02edeb1-77e8-4884-a922-3dc7d608b587" containerName="dnsmasq-dns" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.154056 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a02edeb1-77e8-4884-a922-3dc7d608b587" containerName="dnsmasq-dns" Oct 05 07:15:00 crc kubenswrapper[4935]: E1005 07:15:00.154100 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a02edeb1-77e8-4884-a922-3dc7d608b587" containerName="init" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.154111 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a02edeb1-77e8-4884-a922-3dc7d608b587" containerName="init" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.154420 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a02edeb1-77e8-4884-a922-3dc7d608b587" containerName="dnsmasq-dns" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.155442 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.159737 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.160055 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.170936 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz"] Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.240998 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-config-volume\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.241043 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhwql\" (UniqueName: \"kubernetes.io/projected/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-kube-api-access-nhwql\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.241365 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-secret-volume\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.343658 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-config-volume\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.343715 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhwql\" (UniqueName: \"kubernetes.io/projected/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-kube-api-access-nhwql\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.343787 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-secret-volume\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.346507 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-config-volume\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.360098 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-secret-volume\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.361710 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhwql\" (UniqueName: \"kubernetes.io/projected/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-kube-api-access-nhwql\") pod \"collect-profiles-29327475-gkbbz\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.485681 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:00 crc kubenswrapper[4935]: I1005 07:15:00.968609 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz"] Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.175946 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.266273 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-config-data\") pod \"fbda9936-7e2e-49ec-ad44-29f6451a266b\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.266340 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw4zh\" (UniqueName: \"kubernetes.io/projected/fbda9936-7e2e-49ec-ad44-29f6451a266b-kube-api-access-fw4zh\") pod \"fbda9936-7e2e-49ec-ad44-29f6451a266b\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.266443 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-scripts\") pod \"fbda9936-7e2e-49ec-ad44-29f6451a266b\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.266514 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-combined-ca-bundle\") pod \"fbda9936-7e2e-49ec-ad44-29f6451a266b\" (UID: \"fbda9936-7e2e-49ec-ad44-29f6451a266b\") " Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.272308 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbda9936-7e2e-49ec-ad44-29f6451a266b-kube-api-access-fw4zh" (OuterVolumeSpecName: "kube-api-access-fw4zh") pod "fbda9936-7e2e-49ec-ad44-29f6451a266b" (UID: "fbda9936-7e2e-49ec-ad44-29f6451a266b"). InnerVolumeSpecName "kube-api-access-fw4zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.272428 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-scripts" (OuterVolumeSpecName: "scripts") pod "fbda9936-7e2e-49ec-ad44-29f6451a266b" (UID: "fbda9936-7e2e-49ec-ad44-29f6451a266b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.305853 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbda9936-7e2e-49ec-ad44-29f6451a266b" (UID: "fbda9936-7e2e-49ec-ad44-29f6451a266b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.305961 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-config-data" (OuterVolumeSpecName: "config-data") pod "fbda9936-7e2e-49ec-ad44-29f6451a266b" (UID: "fbda9936-7e2e-49ec-ad44-29f6451a266b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.379157 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.379190 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.379201 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbda9936-7e2e-49ec-ad44-29f6451a266b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.379211 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw4zh\" (UniqueName: \"kubernetes.io/projected/fbda9936-7e2e-49ec-ad44-29f6451a266b-kube-api-access-fw4zh\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.932603 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bm525" event={"ID":"fbda9936-7e2e-49ec-ad44-29f6451a266b","Type":"ContainerDied","Data":"785cb990cbcda9b1211e614b1551e11272ec279717ac97e3c3ffb24982e54750"} Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.932654 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785cb990cbcda9b1211e614b1551e11272ec279717ac97e3c3ffb24982e54750" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.932736 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bm525" Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.940015 4935 generic.go:334] "Generic (PLEG): container finished" podID="20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a" containerID="e5dc00576705b6f955a15b49132057c1684260abcf8020c2a1cb6e30356b9f6a" exitCode=0 Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.940051 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" event={"ID":"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a","Type":"ContainerDied","Data":"e5dc00576705b6f955a15b49132057c1684260abcf8020c2a1cb6e30356b9f6a"} Oct 05 07:15:01 crc kubenswrapper[4935]: I1005 07:15:01.940076 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" event={"ID":"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a","Type":"ContainerStarted","Data":"4b44dfd5a194a1f2c5b30788ec3878d8b9bc6a077b8e7e4b841efc25281745e1"} Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.066725 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.067010 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerName="nova-api-log" containerID="cri-o://42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da" gracePeriod=30 Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.067255 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerName="nova-api-api" containerID="cri-o://49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2" gracePeriod=30 Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.087402 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.087651 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5fa41a7b-d580-42c7-b7ce-55f1799b3dab" containerName="nova-scheduler-scheduler" containerID="cri-o://00055eed91603d646bb93620f446eb6fdaa145639403015e47a6633ffda568ee" gracePeriod=30 Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.100075 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.100397 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-metadata" containerID="cri-o://8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97" gracePeriod=30 Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.100356 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-log" containerID="cri-o://a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a" gracePeriod=30 Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.647485 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.810096 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-config-data\") pod \"78323dc0-73ec-47d4-bdf0-639a3f308db1\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.810205 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-public-tls-certs\") pod \"78323dc0-73ec-47d4-bdf0-639a3f308db1\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.810274 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-internal-tls-certs\") pod \"78323dc0-73ec-47d4-bdf0-639a3f308db1\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.810406 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gjb2\" (UniqueName: \"kubernetes.io/projected/78323dc0-73ec-47d4-bdf0-639a3f308db1-kube-api-access-9gjb2\") pod \"78323dc0-73ec-47d4-bdf0-639a3f308db1\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.810440 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78323dc0-73ec-47d4-bdf0-639a3f308db1-logs\") pod \"78323dc0-73ec-47d4-bdf0-639a3f308db1\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.810481 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-combined-ca-bundle\") pod \"78323dc0-73ec-47d4-bdf0-639a3f308db1\" (UID: \"78323dc0-73ec-47d4-bdf0-639a3f308db1\") " Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.810784 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78323dc0-73ec-47d4-bdf0-639a3f308db1-logs" (OuterVolumeSpecName: "logs") pod "78323dc0-73ec-47d4-bdf0-639a3f308db1" (UID: "78323dc0-73ec-47d4-bdf0-639a3f308db1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.811094 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78323dc0-73ec-47d4-bdf0-639a3f308db1-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.815198 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78323dc0-73ec-47d4-bdf0-639a3f308db1-kube-api-access-9gjb2" (OuterVolumeSpecName: "kube-api-access-9gjb2") pod "78323dc0-73ec-47d4-bdf0-639a3f308db1" (UID: "78323dc0-73ec-47d4-bdf0-639a3f308db1"). InnerVolumeSpecName "kube-api-access-9gjb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.842038 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-config-data" (OuterVolumeSpecName: "config-data") pod "78323dc0-73ec-47d4-bdf0-639a3f308db1" (UID: "78323dc0-73ec-47d4-bdf0-639a3f308db1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.842586 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78323dc0-73ec-47d4-bdf0-639a3f308db1" (UID: "78323dc0-73ec-47d4-bdf0-639a3f308db1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.879984 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "78323dc0-73ec-47d4-bdf0-639a3f308db1" (UID: "78323dc0-73ec-47d4-bdf0-639a3f308db1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.885032 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "78323dc0-73ec-47d4-bdf0-639a3f308db1" (UID: "78323dc0-73ec-47d4-bdf0-639a3f308db1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.913330 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.913361 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.913371 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.913380 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gjb2\" (UniqueName: \"kubernetes.io/projected/78323dc0-73ec-47d4-bdf0-639a3f308db1-kube-api-access-9gjb2\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.913389 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78323dc0-73ec-47d4-bdf0-639a3f308db1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.950518 4935 generic.go:334] "Generic (PLEG): container finished" podID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerID="a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a" exitCode=143 Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.950580 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9400b67d-c149-43bb-b020-e3d2d93f32c1","Type":"ContainerDied","Data":"a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a"} Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.953981 4935 generic.go:334] "Generic (PLEG): container finished" podID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerID="49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2" exitCode=0 Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.954006 4935 generic.go:334] "Generic (PLEG): container finished" podID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerID="42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da" exitCode=143 Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.954353 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.957671 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78323dc0-73ec-47d4-bdf0-639a3f308db1","Type":"ContainerDied","Data":"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2"} Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.957733 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78323dc0-73ec-47d4-bdf0-639a3f308db1","Type":"ContainerDied","Data":"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da"} Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.957751 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78323dc0-73ec-47d4-bdf0-639a3f308db1","Type":"ContainerDied","Data":"f1f9b8b1fd64c05287d4d21a006676608b0375909ca8454e32a4b8de48a4e9cd"} Oct 05 07:15:02 crc kubenswrapper[4935]: I1005 07:15:02.957776 4935 scope.go:117] "RemoveContainer" containerID="49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.002046 4935 scope.go:117] "RemoveContainer" containerID="42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.002206 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.016880 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.036302 4935 scope.go:117] "RemoveContainer" containerID="49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2" Oct 05 07:15:03 crc kubenswrapper[4935]: E1005 07:15:03.036769 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2\": container with ID starting with 49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2 not found: ID does not exist" containerID="49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.036809 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2"} err="failed to get container status \"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2\": rpc error: code = NotFound desc = could not find container \"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2\": container with ID starting with 49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2 not found: ID does not exist" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.036838 4935 scope.go:117] "RemoveContainer" containerID="42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da" Oct 05 07:15:03 crc kubenswrapper[4935]: E1005 07:15:03.037101 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da\": container with ID starting with 42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da not found: ID does not exist" containerID="42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037135 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da"} err="failed to get container status \"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da\": rpc error: code = NotFound desc = could not find container \"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da\": container with ID starting with 42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da not found: ID does not exist" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037159 4935 scope.go:117] "RemoveContainer" containerID="49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037275 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037342 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2"} err="failed to get container status \"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2\": rpc error: code = NotFound desc = could not find container \"49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2\": container with ID starting with 49f05bb25914a6c1c04772775d1cb502cb5d29516dc1a9deda910cf005d809d2 not found: ID does not exist" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037358 4935 scope.go:117] "RemoveContainer" containerID="42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037554 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da"} err="failed to get container status \"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da\": rpc error: code = NotFound desc = could not find container \"42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da\": container with ID starting with 42b1a0bdf61e6ffee3cad6e96a4eedc8efdf002da8c0818e13090ac000dde1da not found: ID does not exist" Oct 05 07:15:03 crc kubenswrapper[4935]: E1005 07:15:03.037851 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerName="nova-api-log" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037869 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerName="nova-api-log" Oct 05 07:15:03 crc kubenswrapper[4935]: E1005 07:15:03.037925 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerName="nova-api-api" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037932 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerName="nova-api-api" Oct 05 07:15:03 crc kubenswrapper[4935]: E1005 07:15:03.037938 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbda9936-7e2e-49ec-ad44-29f6451a266b" containerName="nova-manage" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.037944 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbda9936-7e2e-49ec-ad44-29f6451a266b" containerName="nova-manage" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.038101 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerName="nova-api-log" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.038117 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" containerName="nova-api-api" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.038130 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbda9936-7e2e-49ec-ad44-29f6451a266b" containerName="nova-manage" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.039357 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.041952 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.042125 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.042234 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.047775 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.117315 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf4f5\" (UniqueName: \"kubernetes.io/projected/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-kube-api-access-bf4f5\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.117385 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.117425 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-logs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.117446 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-config-data\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.117499 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.117554 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-public-tls-certs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.219073 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.219125 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-logs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.219200 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-config-data\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.219256 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.219302 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-public-tls-certs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.219337 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf4f5\" (UniqueName: \"kubernetes.io/projected/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-kube-api-access-bf4f5\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.219878 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-logs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.223011 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.223218 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-config-data\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.225351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-public-tls-certs\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.231592 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.233325 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf4f5\" (UniqueName: \"kubernetes.io/projected/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-kube-api-access-bf4f5\") pod \"nova-api-0\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.292317 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.363233 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.422371 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-secret-volume\") pod \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.422580 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhwql\" (UniqueName: \"kubernetes.io/projected/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-kube-api-access-nhwql\") pod \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.422811 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-config-volume\") pod \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\" (UID: \"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a\") " Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.423626 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-config-volume" (OuterVolumeSpecName: "config-volume") pod "20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a" (UID: "20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.424196 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.427210 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a" (UID: "20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.427295 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-kube-api-access-nhwql" (OuterVolumeSpecName: "kube-api-access-nhwql") pod "20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a" (UID: "20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a"). InnerVolumeSpecName "kube-api-access-nhwql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.526094 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhwql\" (UniqueName: \"kubernetes.io/projected/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-kube-api-access-nhwql\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.526386 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.975126 4935 generic.go:334] "Generic (PLEG): container finished" podID="5fa41a7b-d580-42c7-b7ce-55f1799b3dab" containerID="00055eed91603d646bb93620f446eb6fdaa145639403015e47a6633ffda568ee" exitCode=0 Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.975225 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5fa41a7b-d580-42c7-b7ce-55f1799b3dab","Type":"ContainerDied","Data":"00055eed91603d646bb93620f446eb6fdaa145639403015e47a6633ffda568ee"} Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.977721 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.977722 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz" event={"ID":"20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a","Type":"ContainerDied","Data":"4b44dfd5a194a1f2c5b30788ec3878d8b9bc6a077b8e7e4b841efc25281745e1"} Oct 05 07:15:03 crc kubenswrapper[4935]: I1005 07:15:03.977842 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b44dfd5a194a1f2c5b30788ec3878d8b9bc6a077b8e7e4b841efc25281745e1" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.143540 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.228215 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.239685 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-combined-ca-bundle\") pod \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.240013 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6cfv\" (UniqueName: \"kubernetes.io/projected/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-kube-api-access-z6cfv\") pod \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.240129 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-config-data\") pod \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\" (UID: \"5fa41a7b-d580-42c7-b7ce-55f1799b3dab\") " Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.255833 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-kube-api-access-z6cfv" (OuterVolumeSpecName: "kube-api-access-z6cfv") pod "5fa41a7b-d580-42c7-b7ce-55f1799b3dab" (UID: "5fa41a7b-d580-42c7-b7ce-55f1799b3dab"). InnerVolumeSpecName "kube-api-access-z6cfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.278172 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-config-data" (OuterVolumeSpecName: "config-data") pod "5fa41a7b-d580-42c7-b7ce-55f1799b3dab" (UID: "5fa41a7b-d580-42c7-b7ce-55f1799b3dab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.287326 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fa41a7b-d580-42c7-b7ce-55f1799b3dab" (UID: "5fa41a7b-d580-42c7-b7ce-55f1799b3dab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.342866 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.342924 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.342942 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6cfv\" (UniqueName: \"kubernetes.io/projected/5fa41a7b-d580-42c7-b7ce-55f1799b3dab-kube-api-access-z6cfv\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.795412 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78323dc0-73ec-47d4-bdf0-639a3f308db1" path="/var/lib/kubelet/pods/78323dc0-73ec-47d4-bdf0-639a3f308db1/volumes" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.991047 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5fa41a7b-d580-42c7-b7ce-55f1799b3dab","Type":"ContainerDied","Data":"872f15980ca731abcc238bb1cc2e15c97781a8822f5acb47756c7f7b124a43ce"} Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.991084 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.991111 4935 scope.go:117] "RemoveContainer" containerID="00055eed91603d646bb93620f446eb6fdaa145639403015e47a6633ffda568ee" Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.995838 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c","Type":"ContainerStarted","Data":"1ca4b6cdb5511dcb0a9782ff53d490a4721c69f45e3fc5128541c3b64c51ddd3"} Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.995917 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c","Type":"ContainerStarted","Data":"3e0b1900e020511c33b937275937ce79dc33f680831c304611354b614ac0d6a6"} Oct 05 07:15:04 crc kubenswrapper[4935]: I1005 07:15:04.995929 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c","Type":"ContainerStarted","Data":"ce69b9b67f63c1eabab425a4a92471582bea87b0eac36b58ede89a2347a51ed3"} Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.019477 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.029338 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.040991 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:05 crc kubenswrapper[4935]: E1005 07:15:05.041510 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa41a7b-d580-42c7-b7ce-55f1799b3dab" containerName="nova-scheduler-scheduler" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.041541 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa41a7b-d580-42c7-b7ce-55f1799b3dab" containerName="nova-scheduler-scheduler" Oct 05 07:15:05 crc kubenswrapper[4935]: E1005 07:15:05.041595 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a" containerName="collect-profiles" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.041604 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a" containerName="collect-profiles" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.041602 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.041585571 podStartE2EDuration="3.041585571s" podCreationTimestamp="2025-10-05 07:15:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:15:05.034228217 +0000 UTC m=+1338.916854667" watchObservedRunningTime="2025-10-05 07:15:05.041585571 +0000 UTC m=+1338.924212041" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.041838 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a" containerName="collect-profiles" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.041872 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fa41a7b-d580-42c7-b7ce-55f1799b3dab" containerName="nova-scheduler-scheduler" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.042713 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.044656 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.070049 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.207635 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.207695 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cw9j\" (UniqueName: \"kubernetes.io/projected/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-kube-api-access-7cw9j\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.207751 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-config-data\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.246642 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:46962->10.217.0.189:8775: read: connection reset by peer" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.246744 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:46964->10.217.0.189:8775: read: connection reset by peer" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.309436 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cw9j\" (UniqueName: \"kubernetes.io/projected/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-kube-api-access-7cw9j\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.309513 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-config-data\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.309630 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.315058 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.318601 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-config-data\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.330249 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cw9j\" (UniqueName: \"kubernetes.io/projected/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-kube-api-access-7cw9j\") pod \"nova-scheduler-0\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.407564 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.730095 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:15:05 crc kubenswrapper[4935]: W1005 07:15:05.894305 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6d217a2_8ab3_46a8_a70b_58cb5d5cad32.slice/crio-76c1107eef30a0a722711178233696fe10f16a1b12c802573671e8326b03cc7e WatchSource:0}: Error finding container 76c1107eef30a0a722711178233696fe10f16a1b12c802573671e8326b03cc7e: Status 404 returned error can't find the container with id 76c1107eef30a0a722711178233696fe10f16a1b12c802573671e8326b03cc7e Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.896626 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.920936 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-combined-ca-bundle\") pod \"9400b67d-c149-43bb-b020-e3d2d93f32c1\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.921327 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-nova-metadata-tls-certs\") pod \"9400b67d-c149-43bb-b020-e3d2d93f32c1\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.921375 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dpz5\" (UniqueName: \"kubernetes.io/projected/9400b67d-c149-43bb-b020-e3d2d93f32c1-kube-api-access-7dpz5\") pod \"9400b67d-c149-43bb-b020-e3d2d93f32c1\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.921685 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9400b67d-c149-43bb-b020-e3d2d93f32c1-logs\") pod \"9400b67d-c149-43bb-b020-e3d2d93f32c1\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.921980 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-config-data\") pod \"9400b67d-c149-43bb-b020-e3d2d93f32c1\" (UID: \"9400b67d-c149-43bb-b020-e3d2d93f32c1\") " Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.922296 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9400b67d-c149-43bb-b020-e3d2d93f32c1-logs" (OuterVolumeSpecName: "logs") pod "9400b67d-c149-43bb-b020-e3d2d93f32c1" (UID: "9400b67d-c149-43bb-b020-e3d2d93f32c1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.922934 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9400b67d-c149-43bb-b020-e3d2d93f32c1-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.925780 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9400b67d-c149-43bb-b020-e3d2d93f32c1-kube-api-access-7dpz5" (OuterVolumeSpecName: "kube-api-access-7dpz5") pod "9400b67d-c149-43bb-b020-e3d2d93f32c1" (UID: "9400b67d-c149-43bb-b020-e3d2d93f32c1"). InnerVolumeSpecName "kube-api-access-7dpz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.952471 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-config-data" (OuterVolumeSpecName: "config-data") pod "9400b67d-c149-43bb-b020-e3d2d93f32c1" (UID: "9400b67d-c149-43bb-b020-e3d2d93f32c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.953827 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9400b67d-c149-43bb-b020-e3d2d93f32c1" (UID: "9400b67d-c149-43bb-b020-e3d2d93f32c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:05 crc kubenswrapper[4935]: I1005 07:15:05.973190 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9400b67d-c149-43bb-b020-e3d2d93f32c1" (UID: "9400b67d-c149-43bb-b020-e3d2d93f32c1"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.010871 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32","Type":"ContainerStarted","Data":"76c1107eef30a0a722711178233696fe10f16a1b12c802573671e8326b03cc7e"} Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.014480 4935 generic.go:334] "Generic (PLEG): container finished" podID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerID="8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97" exitCode=0 Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.015249 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9400b67d-c149-43bb-b020-e3d2d93f32c1","Type":"ContainerDied","Data":"8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97"} Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.015189 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.015348 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9400b67d-c149-43bb-b020-e3d2d93f32c1","Type":"ContainerDied","Data":"a0acc354536afa4c89bc273bbba8dc40549cd0d35e89d7f3ad2dbae96123e18f"} Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.015388 4935 scope.go:117] "RemoveContainer" containerID="8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.024395 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.024443 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.024461 4935 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9400b67d-c149-43bb-b020-e3d2d93f32c1-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.024475 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dpz5\" (UniqueName: \"kubernetes.io/projected/9400b67d-c149-43bb-b020-e3d2d93f32c1-kube-api-access-7dpz5\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.075183 4935 scope.go:117] "RemoveContainer" containerID="a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.085666 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.093773 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.102250 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:06 crc kubenswrapper[4935]: E1005 07:15:06.102852 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-log" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.102870 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-log" Oct 05 07:15:06 crc kubenswrapper[4935]: E1005 07:15:06.102912 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-metadata" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.102921 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-metadata" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.103163 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-log" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.103190 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" containerName="nova-metadata-metadata" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.105332 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.107882 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.108275 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.112155 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.113342 4935 scope.go:117] "RemoveContainer" containerID="8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97" Oct 05 07:15:06 crc kubenswrapper[4935]: E1005 07:15:06.114874 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97\": container with ID starting with 8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97 not found: ID does not exist" containerID="8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.114940 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97"} err="failed to get container status \"8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97\": rpc error: code = NotFound desc = could not find container \"8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97\": container with ID starting with 8a55f87673c2ae5b83bf0896f5e0fa5206e2b4e856593b1f0c838bb89356ae97 not found: ID does not exist" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.114969 4935 scope.go:117] "RemoveContainer" containerID="a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a" Oct 05 07:15:06 crc kubenswrapper[4935]: E1005 07:15:06.115423 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a\": container with ID starting with a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a not found: ID does not exist" containerID="a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.115451 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a"} err="failed to get container status \"a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a\": rpc error: code = NotFound desc = could not find container \"a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a\": container with ID starting with a3157b652e79a9a2ab8e3ea4a3564c5fc87077dd71c2a2db1c0e979f04d3ab2a not found: ID does not exist" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.228790 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.229280 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1eb003-e111-4a86-b6c7-a899aa0426a0-logs\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.229364 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtgv9\" (UniqueName: \"kubernetes.io/projected/9f1eb003-e111-4a86-b6c7-a899aa0426a0-kube-api-access-vtgv9\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.229401 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.229456 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-config-data\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.330544 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.330618 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1eb003-e111-4a86-b6c7-a899aa0426a0-logs\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.330679 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtgv9\" (UniqueName: \"kubernetes.io/projected/9f1eb003-e111-4a86-b6c7-a899aa0426a0-kube-api-access-vtgv9\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.330712 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.330756 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-config-data\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.331101 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1eb003-e111-4a86-b6c7-a899aa0426a0-logs\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.334114 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-config-data\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.334277 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.334694 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.349384 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtgv9\" (UniqueName: \"kubernetes.io/projected/9f1eb003-e111-4a86-b6c7-a899aa0426a0-kube-api-access-vtgv9\") pod \"nova-metadata-0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.430982 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.792458 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa41a7b-d580-42c7-b7ce-55f1799b3dab" path="/var/lib/kubelet/pods/5fa41a7b-d580-42c7-b7ce-55f1799b3dab/volumes" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.793964 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9400b67d-c149-43bb-b020-e3d2d93f32c1" path="/var/lib/kubelet/pods/9400b67d-c149-43bb-b020-e3d2d93f32c1/volumes" Oct 05 07:15:06 crc kubenswrapper[4935]: I1005 07:15:06.872842 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:07 crc kubenswrapper[4935]: I1005 07:15:07.025668 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32","Type":"ContainerStarted","Data":"2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f"} Oct 05 07:15:07 crc kubenswrapper[4935]: I1005 07:15:07.027276 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f1eb003-e111-4a86-b6c7-a899aa0426a0","Type":"ContainerStarted","Data":"a26d2a64080f031f3f81dc5b5a40e927c359a81f36fa092010bad367d21a0dc1"} Oct 05 07:15:07 crc kubenswrapper[4935]: I1005 07:15:07.061831 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.061799841 podStartE2EDuration="2.061799841s" podCreationTimestamp="2025-10-05 07:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:15:07.046797834 +0000 UTC m=+1340.929424314" watchObservedRunningTime="2025-10-05 07:15:07.061799841 +0000 UTC m=+1340.944426331" Oct 05 07:15:08 crc kubenswrapper[4935]: I1005 07:15:08.040619 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f1eb003-e111-4a86-b6c7-a899aa0426a0","Type":"ContainerStarted","Data":"573548ed93406199c78ab252ce72ca55ded60b3bb8509475fa7af985c3e26a02"} Oct 05 07:15:08 crc kubenswrapper[4935]: I1005 07:15:08.040981 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f1eb003-e111-4a86-b6c7-a899aa0426a0","Type":"ContainerStarted","Data":"122b7e3178e70d3a5f9d4ca95b47f2e8787d8b3c771046206b29a4dc71f41a66"} Oct 05 07:15:08 crc kubenswrapper[4935]: I1005 07:15:08.070066 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.070037667 podStartE2EDuration="2.070037667s" podCreationTimestamp="2025-10-05 07:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:15:08.055075901 +0000 UTC m=+1341.937702371" watchObservedRunningTime="2025-10-05 07:15:08.070037667 +0000 UTC m=+1341.952664167" Oct 05 07:15:10 crc kubenswrapper[4935]: I1005 07:15:10.408771 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 07:15:11 crc kubenswrapper[4935]: I1005 07:15:11.431209 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:15:11 crc kubenswrapper[4935]: I1005 07:15:11.431400 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:15:13 crc kubenswrapper[4935]: I1005 07:15:13.364081 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:15:13 crc kubenswrapper[4935]: I1005 07:15:13.364139 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:15:14 crc kubenswrapper[4935]: I1005 07:15:14.377106 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:15:14 crc kubenswrapper[4935]: I1005 07:15:14.377140 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:15:15 crc kubenswrapper[4935]: I1005 07:15:15.408737 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 07:15:15 crc kubenswrapper[4935]: I1005 07:15:15.433552 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 07:15:16 crc kubenswrapper[4935]: I1005 07:15:16.171014 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 07:15:16 crc kubenswrapper[4935]: I1005 07:15:16.431929 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 07:15:16 crc kubenswrapper[4935]: I1005 07:15:16.431981 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 07:15:17 crc kubenswrapper[4935]: I1005 07:15:17.447012 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:15:17 crc kubenswrapper[4935]: I1005 07:15:17.447036 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:15:23 crc kubenswrapper[4935]: I1005 07:15:23.295337 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 07:15:23 crc kubenswrapper[4935]: I1005 07:15:23.375809 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 07:15:23 crc kubenswrapper[4935]: I1005 07:15:23.376458 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 07:15:23 crc kubenswrapper[4935]: I1005 07:15:23.377332 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 07:15:23 crc kubenswrapper[4935]: I1005 07:15:23.388499 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 07:15:24 crc kubenswrapper[4935]: I1005 07:15:24.222949 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 07:15:24 crc kubenswrapper[4935]: I1005 07:15:24.231709 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 07:15:26 crc kubenswrapper[4935]: I1005 07:15:26.439492 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 07:15:26 crc kubenswrapper[4935]: I1005 07:15:26.443086 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 07:15:26 crc kubenswrapper[4935]: I1005 07:15:26.449002 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 07:15:27 crc kubenswrapper[4935]: I1005 07:15:27.266204 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.025909 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.026663 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerName="cinder-scheduler" containerID="cri-o://24b5a0ee97d3993b21bd1da36e2ba46a94d38c77257bf5f70b85ea35748a294c" gracePeriod=30 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.026797 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerName="probe" containerID="cri-o://c21841257bca2cda12233fe43bab1edd250a07081b7783d9448defce60d379a8" gracePeriod=30 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.061374 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.061647 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api-log" containerID="cri-o://afd975882d932c2ed56cbd1069d50977a2d2abf98e1305a15671c213fc965de8" gracePeriod=30 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.061788 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api" containerID="cri-o://7378ebd5875058e6928f9504c6ce673a929f0b593e3c881ba94728592aaa6422" gracePeriod=30 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.220882 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementdd1f-account-delete-cgh8g"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.222264 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementdd1f-account-delete-cgh8g" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.253704 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementdd1f-account-delete-cgh8g"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.269048 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.310450 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znz9r\" (UniqueName: \"kubernetes.io/projected/0095740d-765a-4cba-b0c8-d36592fad800-kube-api-access-znz9r\") pod \"placementdd1f-account-delete-cgh8g\" (UID: \"0095740d-765a-4cba-b0c8-d36592fad800\") " pod="openstack/placementdd1f-account-delete-cgh8g" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.374966 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.375395 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="abe76393-92f2-4164-a549-5f75ced24fb3" containerName="openstackclient" containerID="cri-o://1d9ad091dc634c8651d67b5d664e7fdae2e562baf23016a53943168af716e873" gracePeriod=2 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.391235 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.424037 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znz9r\" (UniqueName: \"kubernetes.io/projected/0095740d-765a-4cba-b0c8-d36592fad800-kube-api-access-znz9r\") pod \"placementdd1f-account-delete-cgh8g\" (UID: \"0095740d-765a-4cba-b0c8-d36592fad800\") " pod="openstack/placementdd1f-account-delete-cgh8g" Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.424247 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.424298 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data podName:8dcf22e2-2acd-4b49-a2bc-813a50f89409 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:48.924281879 +0000 UTC m=+1382.806908329 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data") pod "rabbitmq-server-0" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409") : configmap "rabbitmq-config-data" not found Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.468978 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.469799 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerName="openstack-network-exporter" containerID="cri-o://453dce919ffa6f7b445b83173fc85bcadcbbe96d524dbe18eb42e012aa8b54c1" gracePeriod=300 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.508663 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znz9r\" (UniqueName: \"kubernetes.io/projected/0095740d-765a-4cba-b0c8-d36592fad800-kube-api-access-znz9r\") pod \"placementdd1f-account-delete-cgh8g\" (UID: \"0095740d-765a-4cba-b0c8-d36592fad800\") " pod="openstack/placementdd1f-account-delete-cgh8g" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.517176 4935 generic.go:334] "Generic (PLEG): container finished" podID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerID="afd975882d932c2ed56cbd1069d50977a2d2abf98e1305a15671c213fc965de8" exitCode=143 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.517427 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7c7acc7a-797f-46cd-b298-1f28fd2951d6","Type":"ContainerDied","Data":"afd975882d932c2ed56cbd1069d50977a2d2abf98e1305a15671c213fc965de8"} Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.521380 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronb3bb-account-delete-v2qzp"] Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.521789 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe76393-92f2-4164-a549-5f75ced24fb3" containerName="openstackclient" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.521802 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe76393-92f2-4164-a549-5f75ced24fb3" containerName="openstackclient" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.522039 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="abe76393-92f2-4164-a549-5f75ced24fb3" containerName="openstackclient" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.522665 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb3bb-account-delete-v2qzp" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.561995 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementdd1f-account-delete-cgh8g" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.572821 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronb3bb-account-delete-v2qzp"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.636602 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-2vfdz"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.644868 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7fqw\" (UniqueName: \"kubernetes.io/projected/d6ccd5ed-23d9-4d0b-99fb-68655c0199b5-kube-api-access-l7fqw\") pod \"neutronb3bb-account-delete-v2qzp\" (UID: \"d6ccd5ed-23d9-4d0b-99fb-68655c0199b5\") " pod="openstack/neutronb3bb-account-delete-v2qzp" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.656468 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-2vfdz"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.718293 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.718556 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="ovn-northd" containerID="cri-o://1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" gracePeriod=30 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.719027 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="openstack-network-exporter" containerID="cri-o://71a780f032841e0e4778281485674a6786913a58eb935c66f913a76d256ba68d" gracePeriod=30 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.749209 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7fqw\" (UniqueName: \"kubernetes.io/projected/d6ccd5ed-23d9-4d0b-99fb-68655c0199b5-kube-api-access-l7fqw\") pod \"neutronb3bb-account-delete-v2qzp\" (UID: \"d6ccd5ed-23d9-4d0b-99fb-68655c0199b5\") " pod="openstack/neutronb3bb-account-delete-v2qzp" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.756781 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-5fwl7"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.766035 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicanb178-account-delete-rnqtn"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.767460 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanb178-account-delete-rnqtn" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.774962 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7fqw\" (UniqueName: \"kubernetes.io/projected/d6ccd5ed-23d9-4d0b-99fb-68655c0199b5-kube-api-access-l7fqw\") pod \"neutronb3bb-account-delete-v2qzp\" (UID: \"d6ccd5ed-23d9-4d0b-99fb-68655c0199b5\") " pod="openstack/neutronb3bb-account-delete-v2qzp" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.797153 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerName="ovsdbserver-sb" containerID="cri-o://d3e5a40a1e26ef3caaba5145609516f5ff29e761ac28f571d0da1f626e989083" gracePeriod=300 Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.829349 4935 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placement-6dbdfd7554-jrf6q" secret="" err="secret \"placement-placement-dockercfg-vbqqd\" not found" Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.853078 4935 secret.go:188] Couldn't get secret openstack/placement-scripts: secret "placement-scripts" not found Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.853348 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts podName:88a7ceab-efbb-4ba4-a5f2-c6702abbc314 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:49.353333191 +0000 UTC m=+1383.235959651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts") pod "placement-6dbdfd7554-jrf6q" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314") : secret "placement-scripts" not found Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.855050 4935 secret.go:188] Couldn't get secret openstack/placement-config-data: secret "placement-config-data" not found Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.855088 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data podName:88a7ceab-efbb-4ba4-a5f2-c6702abbc314 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:49.355077847 +0000 UTC m=+1383.237704307 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data") pod "placement-6dbdfd7554-jrf6q" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314") : secret "placement-config-data" not found Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.955768 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbmhr\" (UniqueName: \"kubernetes.io/projected/b732474b-e717-4d8d-91b7-e00409800f85-kube-api-access-dbmhr\") pod \"barbicanb178-account-delete-rnqtn\" (UID: \"b732474b-e717-4d8d-91b7-e00409800f85\") " pod="openstack/barbicanb178-account-delete-rnqtn" Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.963963 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:15:48 crc kubenswrapper[4935]: E1005 07:15:48.964032 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data podName:8dcf22e2-2acd-4b49-a2bc-813a50f89409 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:49.964013189 +0000 UTC m=+1383.846639649 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data") pod "rabbitmq-server-0" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409") : configmap "rabbitmq-config-data" not found Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.968340 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e808378-4679-42ed-a47c-033e1edfb362" path="/var/lib/kubelet/pods/6e808378-4679-42ed-a47c-033e1edfb362/volumes" Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.968870 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-5fwl7"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.968911 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanb178-account-delete-rnqtn"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.968925 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder9bb0-account-delete-8m44f"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.988848 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder9bb0-account-delete-8m44f"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.989032 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-l5t6l"] Oct 05 07:15:48 crc kubenswrapper[4935]: I1005 07:15:48.989242 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancee563-account-delete-dqlrz"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:48.997876 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder9bb0-account-delete-8m44f" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.002677 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-px6js"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.005557 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancee563-account-delete-dqlrz" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.005788 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-l5t6l" podUID="ca591a2f-54ed-41b1-aa10-be879b218844" containerName="openstack-network-exporter" containerID="cri-o://279993f164ba8d7c189ff4bd349158ae14baabd0b7dea6ef9cb73e2c002dc3b6" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.053082 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb3bb-account-delete-v2qzp" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.069755 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l225h\" (UniqueName: \"kubernetes.io/projected/420a72a6-506a-42e6-b77a-ee5ca8527768-kube-api-access-l225h\") pod \"cinder9bb0-account-delete-8m44f\" (UID: \"420a72a6-506a-42e6-b77a-ee5ca8527768\") " pod="openstack/cinder9bb0-account-delete-8m44f" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.069841 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6rwx\" (UniqueName: \"kubernetes.io/projected/6bd6e7de-59a5-479d-b2b8-7f060883f6c8-kube-api-access-j6rwx\") pod \"glancee563-account-delete-dqlrz\" (UID: \"6bd6e7de-59a5-479d-b2b8-7f060883f6c8\") " pod="openstack/glancee563-account-delete-dqlrz" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.071104 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancee563-account-delete-dqlrz"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.074351 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbmhr\" (UniqueName: \"kubernetes.io/projected/b732474b-e717-4d8d-91b7-e00409800f85-kube-api-access-dbmhr\") pod \"barbicanb178-account-delete-rnqtn\" (UID: \"b732474b-e717-4d8d-91b7-e00409800f85\") " pod="openstack/barbicanb178-account-delete-rnqtn" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.122041 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-nl9vd"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.129630 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-pl4nq"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.144970 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-nl9vd"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.146576 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbmhr\" (UniqueName: \"kubernetes.io/projected/b732474b-e717-4d8d-91b7-e00409800f85-kube-api-access-dbmhr\") pod \"barbicanb178-account-delete-rnqtn\" (UID: \"b732474b-e717-4d8d-91b7-e00409800f85\") " pod="openstack/barbicanb178-account-delete-rnqtn" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.175129 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-n9vr5"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.180629 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6rwx\" (UniqueName: \"kubernetes.io/projected/6bd6e7de-59a5-479d-b2b8-7f060883f6c8-kube-api-access-j6rwx\") pod \"glancee563-account-delete-dqlrz\" (UID: \"6bd6e7de-59a5-479d-b2b8-7f060883f6c8\") " pod="openstack/glancee563-account-delete-dqlrz" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.180817 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l225h\" (UniqueName: \"kubernetes.io/projected/420a72a6-506a-42e6-b77a-ee5ca8527768-kube-api-access-l225h\") pod \"cinder9bb0-account-delete-8m44f\" (UID: \"420a72a6-506a-42e6-b77a-ee5ca8527768\") " pod="openstack/cinder9bb0-account-delete-8m44f" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.226391 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-n9vr5"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.242660 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6rwx\" (UniqueName: \"kubernetes.io/projected/6bd6e7de-59a5-479d-b2b8-7f060883f6c8-kube-api-access-j6rwx\") pod \"glancee563-account-delete-dqlrz\" (UID: \"6bd6e7de-59a5-479d-b2b8-7f060883f6c8\") " pod="openstack/glancee563-account-delete-dqlrz" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.265821 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l225h\" (UniqueName: \"kubernetes.io/projected/420a72a6-506a-42e6-b77a-ee5ca8527768-kube-api-access-l225h\") pod \"cinder9bb0-account-delete-8m44f\" (UID: \"420a72a6-506a-42e6-b77a-ee5ca8527768\") " pod="openstack/cinder9bb0-account-delete-8m44f" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.269484 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell09d1d-account-delete-24dn6"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.284718 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell09d1d-account-delete-24dn6" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.309980 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell09d1d-account-delete-24dn6"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.332881 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanb178-account-delete-rnqtn" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.354090 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.358336 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder9bb0-account-delete-8m44f" Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.396318 4935 secret.go:188] Couldn't get secret openstack/placement-scripts: secret "placement-scripts" not found Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.396395 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts podName:88a7ceab-efbb-4ba4-a5f2-c6702abbc314 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:50.396379479 +0000 UTC m=+1384.279005939 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts") pod "placement-6dbdfd7554-jrf6q" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314") : secret "placement-scripts" not found Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.396689 4935 secret.go:188] Couldn't get secret openstack/placement-config-data: secret "placement-config-data" not found Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.396715 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data podName:88a7ceab-efbb-4ba4-a5f2-c6702abbc314 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:50.396707768 +0000 UTC m=+1384.279334228 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data") pod "placement-6dbdfd7554-jrf6q" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314") : secret "placement-config-data" not found Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.409509 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-mbbgc"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.438063 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancee563-account-delete-dqlrz" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.448996 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-mbbgc"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.497768 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbsh2\" (UniqueName: \"kubernetes.io/projected/7143940e-eedf-48a0-b0dc-4ec91df08ad0-kube-api-access-zbsh2\") pod \"novacell09d1d-account-delete-24dn6\" (UID: \"7143940e-eedf-48a0-b0dc-4ec91df08ad0\") " pod="openstack/novacell09d1d-account-delete-24dn6" Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.501597 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.501685 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data podName:53fe24e1-5efb-4422-ade0-16e315b7e328 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:50.001665214 +0000 UTC m=+1383.884291754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data") pod "rabbitmq-cell1-server-0" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.535701 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi70bb-account-delete-5wwbr"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.540259 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi70bb-account-delete-5wwbr" Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.569429 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d984f8_20c4_4b30_b767_c6106aa279cf.slice/crio-conmon-453dce919ffa6f7b445b83173fc85bcadcbbe96d524dbe18eb42e012aa8b54c1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca591a2f_54ed_41b1_aa10_be879b218844.slice/crio-279993f164ba8d7c189ff4bd349158ae14baabd0b7dea6ef9cb73e2c002dc3b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca591a2f_54ed_41b1_aa10_be879b218844.slice/crio-conmon-279993f164ba8d7c189ff4bd349158ae14baabd0b7dea6ef9cb73e2c002dc3b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d984f8_20c4_4b30_b767_c6106aa279cf.slice/crio-d3e5a40a1e26ef3caaba5145609516f5ff29e761ac28f571d0da1f626e989083.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c412f4e_54aa_4d6c_9606_753f97474c3b.slice/crio-conmon-71a780f032841e0e4778281485674a6786913a58eb935c66f913a76d256ba68d.scope\": RecentStats: unable to find data in memory cache]" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.582808 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi70bb-account-delete-5wwbr"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.599190 4935 generic.go:334] "Generic (PLEG): container finished" podID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerID="71a780f032841e0e4778281485674a6786913a58eb935c66f913a76d256ba68d" exitCode=2 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.599288 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7c412f4e-54aa-4d6c-9606-753f97474c3b","Type":"ContainerDied","Data":"71a780f032841e0e4778281485674a6786913a58eb935c66f913a76d256ba68d"} Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.600984 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbsh2\" (UniqueName: \"kubernetes.io/projected/7143940e-eedf-48a0-b0dc-4ec91df08ad0-kube-api-access-zbsh2\") pod \"novacell09d1d-account-delete-24dn6\" (UID: \"7143940e-eedf-48a0-b0dc-4ec91df08ad0\") " pod="openstack/novacell09d1d-account-delete-24dn6" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.640041 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6dbdfd7554-jrf6q"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.666040 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbsh2\" (UniqueName: \"kubernetes.io/projected/7143940e-eedf-48a0-b0dc-4ec91df08ad0-kube-api-access-zbsh2\") pod \"novacell09d1d-account-delete-24dn6\" (UID: \"7143940e-eedf-48a0-b0dc-4ec91df08ad0\") " pod="openstack/novacell09d1d-account-delete-24dn6" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.680576 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d0d984f8-20c4-4b30-b767-c6106aa279cf/ovsdbserver-sb/0.log" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.680618 4935 generic.go:334] "Generic (PLEG): container finished" podID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerID="453dce919ffa6f7b445b83173fc85bcadcbbe96d524dbe18eb42e012aa8b54c1" exitCode=2 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.680635 4935 generic.go:334] "Generic (PLEG): container finished" podID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerID="d3e5a40a1e26ef3caaba5145609516f5ff29e761ac28f571d0da1f626e989083" exitCode=143 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.680826 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0d984f8-20c4-4b30-b767-c6106aa279cf","Type":"ContainerDied","Data":"453dce919ffa6f7b445b83173fc85bcadcbbe96d524dbe18eb42e012aa8b54c1"} Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.680866 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0d984f8-20c4-4b30-b767-c6106aa279cf","Type":"ContainerDied","Data":"d3e5a40a1e26ef3caaba5145609516f5ff29e761ac28f571d0da1f626e989083"} Oct 05 07:15:49 crc kubenswrapper[4935]: W1005 07:15:49.689112 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0095740d_765a_4cba_b0c8_d36592fad800.slice/crio-30792e5f0d29f0148506ee6aca3f68261e7c621815926d30627de0c32c6af5f9 WatchSource:0}: Error finding container 30792e5f0d29f0148506ee6aca3f68261e7c621815926d30627de0c32c6af5f9: Status 404 returned error can't find the container with id 30792e5f0d29f0148506ee6aca3f68261e7c621815926d30627de0c32c6af5f9 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.694398 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.695077 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="openstack-network-exporter" containerID="cri-o://d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f" gracePeriod=300 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.698379 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell09d1d-account-delete-24dn6" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.703315 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xxfr\" (UniqueName: \"kubernetes.io/projected/819c9cf6-f5e5-46af-9fee-c1c4cc88c29a-kube-api-access-7xxfr\") pod \"novaapi70bb-account-delete-5wwbr\" (UID: \"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a\") " pod="openstack/novaapi70bb-account-delete-5wwbr" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.721855 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-d9xn9"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.782939 4935 generic.go:334] "Generic (PLEG): container finished" podID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerID="c21841257bca2cda12233fe43bab1edd250a07081b7783d9448defce60d379a8" exitCode=0 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.783020 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9acce3fe-5a42-49e7-b3e7-a72f01af5736","Type":"ContainerDied","Data":"c21841257bca2cda12233fe43bab1edd250a07081b7783d9448defce60d379a8"} Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.785620 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-d9xn9"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.790209 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="ovsdbserver-nb" containerID="cri-o://9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" gracePeriod=300 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.808017 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xxfr\" (UniqueName: \"kubernetes.io/projected/819c9cf6-f5e5-46af-9fee-c1c4cc88c29a-kube-api-access-7xxfr\") pod \"novaapi70bb-account-delete-5wwbr\" (UID: \"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a\") " pod="openstack/novaapi70bb-account-delete-5wwbr" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.809493 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-l5t6l_ca591a2f-54ed-41b1-aa10-be879b218844/openstack-network-exporter/0.log" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.809541 4935 generic.go:334] "Generic (PLEG): container finished" podID="ca591a2f-54ed-41b1-aa10-be879b218844" containerID="279993f164ba8d7c189ff4bd349158ae14baabd0b7dea6ef9cb73e2c002dc3b6" exitCode=2 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.809744 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6dbdfd7554-jrf6q" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerName="placement-log" containerID="cri-o://27997ed35ed34c0a762315ae80ed4c65e006d4cd919243a41130b860e5a4e10a" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.809836 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-l5t6l" event={"ID":"ca591a2f-54ed-41b1-aa10-be879b218844","Type":"ContainerDied","Data":"279993f164ba8d7c189ff4bd349158ae14baabd0b7dea6ef9cb73e2c002dc3b6"} Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.810031 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6dbdfd7554-jrf6q" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerName="placement-api" containerID="cri-o://829086067b4c67ac2011dcae6d0f8b4b1d850acfb49e4cf8b466a30f81bb3be2" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.823947 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56bb7fd5-dmzld"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.824154 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56bb7fd5-dmzld" podUID="27a28306-70be-4556-9659-64999f775195" containerName="neutron-api" containerID="cri-o://a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.824292 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56bb7fd5-dmzld" podUID="27a28306-70be-4556-9659-64999f775195" containerName="neutron-httpd" containerID="cri-o://16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.864292 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xxfr\" (UniqueName: \"kubernetes.io/projected/819c9cf6-f5e5-46af-9fee-c1c4cc88c29a-kube-api-access-7xxfr\") pod \"novaapi70bb-account-delete-5wwbr\" (UID: \"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a\") " pod="openstack/novaapi70bb-account-delete-5wwbr" Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.874571 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.875950 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95 is running failed: container process not found" containerID="9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.877312 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-z5h55"] Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.880806 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95 is running failed: container process not found" containerID="9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 05 07:15:49 crc kubenswrapper[4935]: E1005 07:15:49.880865 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="ovsdbserver-nb" Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.902665 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-z5h55"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.950147 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.950719 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-server" containerID="cri-o://a461381cdfc3c2eee74eb1bdcf7d5ebfdb0fb05a83a09c83b9bcd518fb51b2d5" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951162 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="swift-recon-cron" containerID="cri-o://fbfc0f5f81ecb328f21528e6e7fd64b97e302cbe3748716791576b9c52740ca0" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951215 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="rsync" containerID="cri-o://f8b9ec64a993b5fc267da34859f562afb461e51a8c3449b8a4980e4dbec60aab" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951245 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-expirer" containerID="cri-o://b90d88185b137f6f204daabf2f63d9c6ca4190f72a5e20e02ad439fe47b1d9e2" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951302 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-updater" containerID="cri-o://8ae6a7f5c8eb347b93d521a8aa9df7e0b4edee491ce3dd6e4dd29fd1cdc12733" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951336 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-auditor" containerID="cri-o://5811d6a7453e1bdaa6a37fe17b4607303b6b13e3e3d1fe5e319d227f3a5e3efb" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951367 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-replicator" containerID="cri-o://f07908304e5257af7f03f59b8ba6637156f346dca17601807d91da7efcc445e9" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951398 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-server" containerID="cri-o://06e262448da43be1f59f8206d3ca149b097157a5612b8e5d368736aac9a62a57" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951485 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-updater" containerID="cri-o://ba8102ea51bc30c759201a866a8431d1baaf2f6f474cc96b36529ea5a04451ec" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951523 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-auditor" containerID="cri-o://cd237a014093f689d7df188672e570ebe94fb2b0f49e1d1a6c58473d64c3b740" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951553 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-replicator" containerID="cri-o://b18142ab07c004a98bff59b529be5e4c49e5ff54caf77526350e39ce8c20c39c" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951580 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-server" containerID="cri-o://cd51894ecb6651dbb869d73d13d130a6f2fb7d28b0f7980f406450f51c4449b9" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951607 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-reaper" containerID="cri-o://95b99cc19f9b362efc74b2612128d2a9a4247e4bfa40d9b7ca5e1c764c78c5c9" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951644 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-auditor" containerID="cri-o://5fb020c3045a74d5048d67e9e8f939b0d2f9348bab133b9b65f5cfb8ff321504" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.951685 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-replicator" containerID="cri-o://22ba6a3f7d0c659cde8f9b4b899df8e9b5b426c7a3f7e76b86ffc508196c5584" gracePeriod=30 Oct 05 07:15:49 crc kubenswrapper[4935]: I1005 07:15:49.984047 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-dd1f-account-create-tkrh9"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.006001 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-dd1f-account-create-tkrh9"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.019417 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi70bb-account-delete-5wwbr" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.026722 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementdd1f-account-delete-cgh8g"] Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.036352 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.036456 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data podName:53fe24e1-5efb-4422-ade0-16e315b7e328 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:51.036427633 +0000 UTC m=+1384.919054093 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data") pod "rabbitmq-cell1-server-0" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.043570 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.043689 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data podName:8dcf22e2-2acd-4b49-a2bc-813a50f89409 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:52.043665115 +0000 UTC m=+1385.926291575 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data") pod "rabbitmq-server-0" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409") : configmap "rabbitmq-config-data" not found Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.192019 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-l5t6l_ca591a2f-54ed-41b1-aa10-be879b218844/openstack-network-exporter/0.log" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.192192 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.217529 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-4w7kp"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.241470 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-4w7kp"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.268547 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" containerID="cri-o://ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" gracePeriod=29 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.270293 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-metrics-certs-tls-certs\") pod \"ca591a2f-54ed-41b1-aa10-be879b218844\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.270386 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovn-rundir\") pod \"ca591a2f-54ed-41b1-aa10-be879b218844\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.270416 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca591a2f-54ed-41b1-aa10-be879b218844-config\") pod \"ca591a2f-54ed-41b1-aa10-be879b218844\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.270504 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovs-rundir\") pod \"ca591a2f-54ed-41b1-aa10-be879b218844\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.270562 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-combined-ca-bundle\") pod \"ca591a2f-54ed-41b1-aa10-be879b218844\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.270648 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thh58\" (UniqueName: \"kubernetes.io/projected/ca591a2f-54ed-41b1-aa10-be879b218844-kube-api-access-thh58\") pod \"ca591a2f-54ed-41b1-aa10-be879b218844\" (UID: \"ca591a2f-54ed-41b1-aa10-be879b218844\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.273472 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca591a2f-54ed-41b1-aa10-be879b218844-config" (OuterVolumeSpecName: "config") pod "ca591a2f-54ed-41b1-aa10-be879b218844" (UID: "ca591a2f-54ed-41b1-aa10-be879b218844"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.280990 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "ca591a2f-54ed-41b1-aa10-be879b218844" (UID: "ca591a2f-54ed-41b1-aa10-be879b218844"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.281271 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "ca591a2f-54ed-41b1-aa10-be879b218844" (UID: "ca591a2f-54ed-41b1-aa10-be879b218844"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.282790 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca591a2f-54ed-41b1-aa10-be879b218844-kube-api-access-thh58" (OuterVolumeSpecName: "kube-api-access-thh58") pod "ca591a2f-54ed-41b1-aa10-be879b218844" (UID: "ca591a2f-54ed-41b1-aa10-be879b218844"). InnerVolumeSpecName "kube-api-access-thh58". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.292967 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementdd1f-account-delete-cgh8g"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.300936 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-bm525"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.358419 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca591a2f-54ed-41b1-aa10-be879b218844" (UID: "ca591a2f-54ed-41b1-aa10-be879b218844"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.363067 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-bm525"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.373242 4935 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.373267 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.373276 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thh58\" (UniqueName: \"kubernetes.io/projected/ca591a2f-54ed-41b1-aa10-be879b218844-kube-api-access-thh58\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.373286 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ca591a2f-54ed-41b1-aa10-be879b218844-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.373297 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca591a2f-54ed-41b1-aa10-be879b218844-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.390093 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-47z4r"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.390363 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerName="dnsmasq-dns" containerID="cri-o://5c401639335e92358e96e3ee9d7c9cbcdd83e9b6aa688f56a6c57891805ffe6d" gracePeriod=10 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.460058 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.460391 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" containerName="glance-log" containerID="cri-o://83dd3f9b8ccfa6c15a5abedf1cc0dba702eb5c6c024d7e58ef7b5517611fcd66" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.460537 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" containerName="glance-httpd" containerID="cri-o://fc1c1a196142dfa9f9def44b621bd771b1f8fbde0b745b2687d81b122f138212" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.477141 4935 secret.go:188] Couldn't get secret openstack/placement-scripts: secret "placement-scripts" not found Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.477293 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts podName:88a7ceab-efbb-4ba4-a5f2-c6702abbc314 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:52.477273247 +0000 UTC m=+1386.359899717 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts") pod "placement-6dbdfd7554-jrf6q" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314") : secret "placement-scripts" not found Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.477700 4935 secret.go:188] Couldn't get secret openstack/placement-config-data: secret "placement-config-data" not found Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.478060 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data podName:88a7ceab-efbb-4ba4-a5f2-c6702abbc314 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:52.478049478 +0000 UTC m=+1386.360675938 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data") pod "placement-6dbdfd7554-jrf6q" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314") : secret "placement-config-data" not found Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.493368 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.493592 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerName="glance-log" containerID="cri-o://af59bf08d75bada41121a63c5849541082f4196ebb916ac86beff7f230324a9b" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.493962 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerName="glance-httpd" containerID="cri-o://f98ff5241a039270a3b05822fea3c87f18859f74162897e6acb797aeca39838d" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.522677 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-57fd77d677-fr4zj"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.523191 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-57fd77d677-fr4zj" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" containerName="barbican-worker-log" containerID="cri-o://a5103fe4899d57404d97b7ee8251771675c6a622925145cac9caa880ee1b5c91" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.523301 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-57fd77d677-fr4zj" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" containerName="barbican-worker" containerID="cri-o://f1adbd1314cf52e09287126eae2dabba2ce549b2976316b1a565b87702fb1644" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.540485 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74989f54d8-bw287"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.540756 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74989f54d8-bw287" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api-log" containerID="cri-o://9b2b88228c11fcf40bd73c44aadf5f4a93619dfc6b49a5c328edcf98f4d93a7e" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.548698 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74989f54d8-bw287" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api" containerID="cri-o://119888a24ef912962cb413c88d9903d21aad373b3de77a30cfb8ecc482ff8f75" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.562518 4935 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 05 07:15:50 crc kubenswrapper[4935]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 05 07:15:50 crc kubenswrapper[4935]: + source /usr/local/bin/container-scripts/functions Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNBridge=br-int Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNRemote=tcp:localhost:6642 Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNEncapType=geneve Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNAvailabilityZones= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ EnableChassisAsGateway=true Oct 05 07:15:50 crc kubenswrapper[4935]: ++ PhysicalNetworks= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNHostName= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 05 07:15:50 crc kubenswrapper[4935]: ++ ovs_dir=/var/lib/openvswitch Oct 05 07:15:50 crc kubenswrapper[4935]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 05 07:15:50 crc kubenswrapper[4935]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 05 07:15:50 crc kubenswrapper[4935]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + sleep 0.5 Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + sleep 0.5 Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + cleanup_ovsdb_server_semaphore Oct 05 07:15:50 crc kubenswrapper[4935]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:15:50 crc kubenswrapper[4935]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 05 07:15:50 crc kubenswrapper[4935]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-pl4nq" message=< Oct 05 07:15:50 crc kubenswrapper[4935]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 05 07:15:50 crc kubenswrapper[4935]: + source /usr/local/bin/container-scripts/functions Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNBridge=br-int Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNRemote=tcp:localhost:6642 Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNEncapType=geneve Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNAvailabilityZones= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ EnableChassisAsGateway=true Oct 05 07:15:50 crc kubenswrapper[4935]: ++ PhysicalNetworks= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNHostName= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 05 07:15:50 crc kubenswrapper[4935]: ++ ovs_dir=/var/lib/openvswitch Oct 05 07:15:50 crc kubenswrapper[4935]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 05 07:15:50 crc kubenswrapper[4935]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 05 07:15:50 crc kubenswrapper[4935]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + sleep 0.5 Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + sleep 0.5 Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + cleanup_ovsdb_server_semaphore Oct 05 07:15:50 crc kubenswrapper[4935]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:15:50 crc kubenswrapper[4935]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 05 07:15:50 crc kubenswrapper[4935]: > Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.562563 4935 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 05 07:15:50 crc kubenswrapper[4935]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 05 07:15:50 crc kubenswrapper[4935]: + source /usr/local/bin/container-scripts/functions Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNBridge=br-int Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNRemote=tcp:localhost:6642 Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNEncapType=geneve Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNAvailabilityZones= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ EnableChassisAsGateway=true Oct 05 07:15:50 crc kubenswrapper[4935]: ++ PhysicalNetworks= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ OVNHostName= Oct 05 07:15:50 crc kubenswrapper[4935]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 05 07:15:50 crc kubenswrapper[4935]: ++ ovs_dir=/var/lib/openvswitch Oct 05 07:15:50 crc kubenswrapper[4935]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 05 07:15:50 crc kubenswrapper[4935]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 05 07:15:50 crc kubenswrapper[4935]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + sleep 0.5 Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + sleep 0.5 Oct 05 07:15:50 crc kubenswrapper[4935]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:15:50 crc kubenswrapper[4935]: + cleanup_ovsdb_server_semaphore Oct 05 07:15:50 crc kubenswrapper[4935]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:15:50 crc kubenswrapper[4935]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 05 07:15:50 crc kubenswrapper[4935]: > pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" containerID="cri-o://81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.562600 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" containerID="cri-o://81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" gracePeriod=29 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.568058 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-b2q9j"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.586074 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-b2q9j"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.600037 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7cbc6494d6-8n275"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.600293 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerName="barbican-keystone-listener-log" containerID="cri-o://8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.600952 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerName="barbican-keystone-listener" containerID="cri-o://9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.618962 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b3bb-account-create-pddl8"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.643727 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b3bb-account-create-pddl8"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.650655 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "ca591a2f-54ed-41b1-aa10-be879b218844" (UID: "ca591a2f-54ed-41b1-aa10-be879b218844"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.673344 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.674838 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-kpphs"] Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.683166 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.684503 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca591a2f-54ed-41b1-aa10-be879b218844-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.685475 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:15:50 crc kubenswrapper[4935]: E1005 07:15:50.685528 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="ovn-northd" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.695701 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb3bb-account-delete-v2qzp"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.711703 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b178-account-create-tl96t"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.734570 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-kpphs"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.741118 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b178-account-create-tl96t"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.746628 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanb178-account-delete-rnqtn"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.753288 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-5z299"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.759102 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-5z299"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.763778 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9bb0-account-create-g9ffr"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.768468 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d0d984f8-20c4-4b30-b767-c6106aa279cf/ovsdbserver-sb/0.log" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.768558 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.770159 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9bb0-account-create-g9ffr"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.776440 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder9bb0-account-delete-8m44f"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.786407 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgpdp\" (UniqueName: \"kubernetes.io/projected/d0d984f8-20c4-4b30-b767-c6106aa279cf-kube-api-access-sgpdp\") pod \"d0d984f8-20c4-4b30-b767-c6106aa279cf\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.786447 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdb-rundir\") pod \"d0d984f8-20c4-4b30-b767-c6106aa279cf\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.786490 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-combined-ca-bundle\") pod \"d0d984f8-20c4-4b30-b767-c6106aa279cf\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.786612 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"d0d984f8-20c4-4b30-b767-c6106aa279cf\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.786663 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-config\") pod \"d0d984f8-20c4-4b30-b767-c6106aa279cf\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.786682 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-metrics-certs-tls-certs\") pod \"d0d984f8-20c4-4b30-b767-c6106aa279cf\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.786764 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdbserver-sb-tls-certs\") pod \"d0d984f8-20c4-4b30-b767-c6106aa279cf\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.786806 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-scripts\") pod \"d0d984f8-20c4-4b30-b767-c6106aa279cf\" (UID: \"d0d984f8-20c4-4b30-b767-c6106aa279cf\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.788408 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "d0d984f8-20c4-4b30-b767-c6106aa279cf" (UID: "d0d984f8-20c4-4b30-b767-c6106aa279cf"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.788970 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-config" (OuterVolumeSpecName: "config") pod "d0d984f8-20c4-4b30-b767-c6106aa279cf" (UID: "d0d984f8-20c4-4b30-b767-c6106aa279cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.789610 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-scripts" (OuterVolumeSpecName: "scripts") pod "d0d984f8-20c4-4b30-b767-c6106aa279cf" (UID: "d0d984f8-20c4-4b30-b767-c6106aa279cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.804632 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f27b0edc-d2a4-418e-8c62-82f1287cc1b4/ovsdbserver-nb/0.log" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.804706 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.808964 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf18d65-b3e6-41ef-a263-bd8642d272ed" path="/var/lib/kubelet/pods/0bf18d65-b3e6-41ef-a263-bd8642d272ed/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.812730 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="115ec2f7-7585-4a18-916b-7e253535e497" path="/var/lib/kubelet/pods/115ec2f7-7585-4a18-916b-7e253535e497/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.816450 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d984f8-20c4-4b30-b767-c6106aa279cf-kube-api-access-sgpdp" (OuterVolumeSpecName: "kube-api-access-sgpdp") pod "d0d984f8-20c4-4b30-b767-c6106aa279cf" (UID: "d0d984f8-20c4-4b30-b767-c6106aa279cf"). InnerVolumeSpecName "kube-api-access-sgpdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.819530 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "d0d984f8-20c4-4b30-b767-c6106aa279cf" (UID: "d0d984f8-20c4-4b30-b767-c6106aa279cf"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.824924 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d21608a-d4fa-4bba-924f-2f62ff4e8ebd" path="/var/lib/kubelet/pods/1d21608a-d4fa-4bba-924f-2f62ff4e8ebd/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.832427 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="216a3a9f-2a37-4b71-baf5-0dba7511c4e2" path="/var/lib/kubelet/pods/216a3a9f-2a37-4b71-baf5-0dba7511c4e2/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.833278 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2da110a9-7934-4fc0-88dc-8025c46413cb" path="/var/lib/kubelet/pods/2da110a9-7934-4fc0-88dc-8025c46413cb/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.835543 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ef780aa-5f7e-4fa2-a019-6ee2c3993692" path="/var/lib/kubelet/pods/2ef780aa-5f7e-4fa2-a019-6ee2c3993692/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.839978 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42c5057a-3e6a-45fd-978c-6619b1a184ca" path="/var/lib/kubelet/pods/42c5057a-3e6a-45fd-978c-6619b1a184ca/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.840554 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bff7991-7fc7-4034-82dd-bc71fc7b3f7a" path="/var/lib/kubelet/pods/5bff7991-7fc7-4034-82dd-bc71fc7b3f7a/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.848433 4935 generic.go:334] "Generic (PLEG): container finished" podID="0095740d-765a-4cba-b0c8-d36592fad800" containerID="d2d2b4c5eb14625e816f7c37ac0375d97c1c1a9134db1e83ded9f96ed1398e3b" exitCode=0 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.850503 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ef5a833-c6c9-43c7-ba5c-92b335d7ace7" path="/var/lib/kubelet/pods/6ef5a833-c6c9-43c7-ba5c-92b335d7ace7/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.852200 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="810c5536-9646-48ee-a6ac-04381b8c0cd5" path="/var/lib/kubelet/pods/810c5536-9646-48ee-a6ac-04381b8c0cd5/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.853520 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a278b3a-993f-427f-9ab5-d189cbfb4a69" path="/var/lib/kubelet/pods/8a278b3a-993f-427f-9ab5-d189cbfb4a69/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.854052 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2def690-05d7-4196-b383-58669b274802" path="/var/lib/kubelet/pods/c2def690-05d7-4196-b383-58669b274802/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.875551 4935 generic.go:334] "Generic (PLEG): container finished" podID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerID="5c401639335e92358e96e3ee9d7c9cbcdd83e9b6aa688f56a6c57891805ffe6d" exitCode=0 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.876674 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7a3c603-bb1a-4a34-80e3-4cae0ef2815b" path="/var/lib/kubelet/pods/c7a3c603-bb1a-4a34-80e3-4cae0ef2815b/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.880830 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9ba8cbb-45b9-49e5-b5a0-9277b724b607" path="/var/lib/kubelet/pods/e9ba8cbb-45b9-49e5-b5a0-9277b724b607/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.899290 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbda9936-7e2e-49ec-ad44-29f6451a266b" path="/var/lib/kubelet/pods/fbda9936-7e2e-49ec-ad44-29f6451a266b/volumes" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.908965 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-l5t6l_ca591a2f-54ed-41b1-aa10-be879b218844/openstack-network-exporter/0.log" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.909129 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-l5t6l" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.913738 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0d984f8-20c4-4b30-b767-c6106aa279cf" (UID: "d0d984f8-20c4-4b30-b767-c6106aa279cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.921451 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f27b0edc-d2a4-418e-8c62-82f1287cc1b4/ovsdbserver-nb/0.log" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.921519 4935 generic.go:334] "Generic (PLEG): container finished" podID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerID="d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f" exitCode=2 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.921541 4935 generic.go:334] "Generic (PLEG): container finished" podID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerID="9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" exitCode=143 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.921739 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.930465 4935 generic.go:334] "Generic (PLEG): container finished" podID="abe76393-92f2-4164-a549-5f75ced24fb3" containerID="1d9ad091dc634c8651d67b5d664e7fdae2e562baf23016a53943168af716e873" exitCode=137 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.947436 4935 generic.go:334] "Generic (PLEG): container finished" podID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerID="27997ed35ed34c0a762315ae80ed4c65e006d4cd919243a41130b860e5a4e10a" exitCode=143 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.954611 4935 generic.go:334] "Generic (PLEG): container finished" podID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerID="af59bf08d75bada41121a63c5849541082f4196ebb916ac86beff7f230324a9b" exitCode=143 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.956373 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdbserver-nb-tls-certs\") pod \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.956738 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldxzj\" (UniqueName: \"kubernetes.io/projected/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-kube-api-access-ldxzj\") pod \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.956855 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdb-rundir\") pod \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.956908 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-config\") pod \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.957007 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-combined-ca-bundle\") pod \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.957025 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.957071 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-metrics-certs-tls-certs\") pod \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.957147 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-scripts\") pod \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\" (UID: \"f27b0edc-d2a4-418e-8c62-82f1287cc1b4\") " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.963849 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.963880 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgpdp\" (UniqueName: \"kubernetes.io/projected/d0d984f8-20c4-4b30-b767-c6106aa279cf-kube-api-access-sgpdp\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.963912 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.963923 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.963953 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.963964 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d984f8-20c4-4b30-b767-c6106aa279cf-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966460 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementdd1f-account-delete-cgh8g" event={"ID":"0095740d-765a-4cba-b0c8-d36592fad800","Type":"ContainerDied","Data":"d2d2b4c5eb14625e816f7c37ac0375d97c1c1a9134db1e83ded9f96ed1398e3b"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966501 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementdd1f-account-delete-cgh8g" event={"ID":"0095740d-765a-4cba-b0c8-d36592fad800","Type":"ContainerStarted","Data":"30792e5f0d29f0148506ee6aca3f68261e7c621815926d30627de0c32c6af5f9"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966513 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-bdqj5"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966678 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" event={"ID":"8bf67c05-04fb-414a-aa6e-8cdc3831cb94","Type":"ContainerDied","Data":"5c401639335e92358e96e3ee9d7c9cbcdd83e9b6aa688f56a6c57891805ffe6d"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966693 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-bdqj5"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966708 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-l5t6l" event={"ID":"ca591a2f-54ed-41b1-aa10-be879b218844","Type":"ContainerDied","Data":"d6edad8ea6f25d5359cf33edc983b72f7e48db74dfbbad4954ae2e4a2451ad55"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966728 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancee563-account-delete-dqlrz"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966741 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-e563-account-create-7ttqx"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966750 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f27b0edc-d2a4-418e-8c62-82f1287cc1b4","Type":"ContainerDied","Data":"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966769 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f27b0edc-d2a4-418e-8c62-82f1287cc1b4","Type":"ContainerDied","Data":"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966779 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f27b0edc-d2a4-418e-8c62-82f1287cc1b4","Type":"ContainerDied","Data":"7799e4cd4f8cf427bbdb16723cf47a8cede7eaf5eebd1674cb4e809090bd4dfc"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966800 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-e563-account-create-7ttqx"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966813 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbdfd7554-jrf6q" event={"ID":"88a7ceab-efbb-4ba4-a5f2-c6702abbc314","Type":"ContainerDied","Data":"27997ed35ed34c0a762315ae80ed4c65e006d4cd919243a41130b860e5a4e10a"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966824 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966837 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74e20071-1ff2-4fae-ae8c-beb9dbce35ce","Type":"ContainerDied","Data":"af59bf08d75bada41121a63c5849541082f4196ebb916ac86beff7f230324a9b"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966849 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966868 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-4wf8f"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966879 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-4wf8f"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966906 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1098-account-create-5zdbm"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966918 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1098-account-create-5zdbm"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.966935 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.967425 4935 scope.go:117] "RemoveContainer" containerID="279993f164ba8d7c189ff4bd349158ae14baabd0b7dea6ef9cb73e2c002dc3b6" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.970552 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-log" containerID="cri-o://122b7e3178e70d3a5f9d4ca95b47f2e8787d8b3c771046206b29a4dc71f41a66" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.970808 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-metadata" containerID="cri-o://573548ed93406199c78ab252ce72ca55ded60b3bb8509475fa7af985c3e26a02" gracePeriod=30 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.985556 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-rnhsc"] Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.985803 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "f27b0edc-d2a4-418e-8c62-82f1287cc1b4" (UID: "f27b0edc-d2a4-418e-8c62-82f1287cc1b4"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.992284 4935 generic.go:334] "Generic (PLEG): container finished" podID="6c512e7c-9418-43b4-8235-4d62fa487556" containerID="83dd3f9b8ccfa6c15a5abedf1cc0dba702eb5c6c024d7e58ef7b5517611fcd66" exitCode=143 Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.992421 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c512e7c-9418-43b4-8235-4d62fa487556","Type":"ContainerDied","Data":"83dd3f9b8ccfa6c15a5abedf1cc0dba702eb5c6c024d7e58ef7b5517611fcd66"} Oct 05 07:15:50 crc kubenswrapper[4935]: I1005 07:15:50.999160 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "f27b0edc-d2a4-418e-8c62-82f1287cc1b4" (UID: "f27b0edc-d2a4-418e-8c62-82f1287cc1b4"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.011340 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.011416 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-scripts" (OuterVolumeSpecName: "scripts") pod "f27b0edc-d2a4-418e-8c62-82f1287cc1b4" (UID: "f27b0edc-d2a4-418e-8c62-82f1287cc1b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.044768 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-config" (OuterVolumeSpecName: "config") pod "f27b0edc-d2a4-418e-8c62-82f1287cc1b4" (UID: "f27b0edc-d2a4-418e-8c62-82f1287cc1b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.048371 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-kube-api-access-ldxzj" (OuterVolumeSpecName: "kube-api-access-ldxzj") pod "f27b0edc-d2a4-418e-8c62-82f1287cc1b4" (UID: "f27b0edc-d2a4-418e-8c62-82f1287cc1b4"). InnerVolumeSpecName "kube-api-access-ldxzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.055729 4935 scope.go:117] "RemoveContainer" containerID="d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.065748 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-rnhsc"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.067064 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.067083 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldxzj\" (UniqueName: \"kubernetes.io/projected/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-kube-api-access-ldxzj\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.067092 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.067104 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.067122 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 05 07:15:51 crc kubenswrapper[4935]: E1005 07:15:51.073535 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:51 crc kubenswrapper[4935]: E1005 07:15:51.073600 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data podName:53fe24e1-5efb-4422-ade0-16e315b7e328 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:53.073583084 +0000 UTC m=+1386.956209544 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data") pod "rabbitmq-cell1-server-0" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.075799 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.076079 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-log" containerID="cri-o://3e0b1900e020511c33b937275937ce79dc33f680831c304611354b614ac0d6a6" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.076252 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-api" containerID="cri-o://1ca4b6cdb5511dcb0a9782ff53d490a4721c69f45e3fc5128541c3b64c51ddd3" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.090526 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi70bb-account-delete-5wwbr"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.104482 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.106686 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-70bb-account-create-2nf56"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.113212 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "d0d984f8-20c4-4b30-b767-c6106aa279cf" (UID: "d0d984f8-20c4-4b30-b767-c6106aa279cf"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.125707 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" containerName="rabbitmq" containerID="cri-o://cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a" gracePeriod=604800 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.151697 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-70bb-account-create-2nf56"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.168217 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rtjh\" (UniqueName: \"kubernetes.io/projected/abe76393-92f2-4164-a549-5f75ced24fb3-kube-api-access-8rtjh\") pod \"abe76393-92f2-4164-a549-5f75ced24fb3\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.168312 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config\") pod \"abe76393-92f2-4164-a549-5f75ced24fb3\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.168351 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-combined-ca-bundle\") pod \"abe76393-92f2-4164-a549-5f75ced24fb3\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.168424 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config-secret\") pod \"abe76393-92f2-4164-a549-5f75ced24fb3\" (UID: \"abe76393-92f2-4164-a549-5f75ced24fb3\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169446 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="f8b9ec64a993b5fc267da34859f562afb461e51a8c3449b8a4980e4dbec60aab" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169484 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="b90d88185b137f6f204daabf2f63d9c6ca4190f72a5e20e02ad439fe47b1d9e2" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169495 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="8ae6a7f5c8eb347b93d521a8aa9df7e0b4edee491ce3dd6e4dd29fd1cdc12733" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169504 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="5811d6a7453e1bdaa6a37fe17b4607303b6b13e3e3d1fe5e319d227f3a5e3efb" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169511 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="f07908304e5257af7f03f59b8ba6637156f346dca17601807d91da7efcc445e9" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169517 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="06e262448da43be1f59f8206d3ca149b097157a5612b8e5d368736aac9a62a57" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169523 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="ba8102ea51bc30c759201a866a8431d1baaf2f6f474cc96b36529ea5a04451ec" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169529 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="cd237a014093f689d7df188672e570ebe94fb2b0f49e1d1a6c58473d64c3b740" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169534 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="b18142ab07c004a98bff59b529be5e4c49e5ff54caf77526350e39ce8c20c39c" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169540 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="cd51894ecb6651dbb869d73d13d130a6f2fb7d28b0f7980f406450f51c4449b9" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169546 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="95b99cc19f9b362efc74b2612128d2a9a4247e4bfa40d9b7ca5e1c764c78c5c9" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169553 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="5fb020c3045a74d5048d67e9e8f939b0d2f9348bab133b9b65f5cfb8ff321504" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169560 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="22ba6a3f7d0c659cde8f9b4b899df8e9b5b426c7a3f7e76b86ffc508196c5584" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.169567 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="a461381cdfc3c2eee74eb1bdcf7d5ebfdb0fb05a83a09c83b9bcd518fb51b2d5" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170619 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"f8b9ec64a993b5fc267da34859f562afb461e51a8c3449b8a4980e4dbec60aab"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170672 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"b90d88185b137f6f204daabf2f63d9c6ca4190f72a5e20e02ad439fe47b1d9e2"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170720 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"8ae6a7f5c8eb347b93d521a8aa9df7e0b4edee491ce3dd6e4dd29fd1cdc12733"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170730 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"5811d6a7453e1bdaa6a37fe17b4607303b6b13e3e3d1fe5e319d227f3a5e3efb"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170739 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"f07908304e5257af7f03f59b8ba6637156f346dca17601807d91da7efcc445e9"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170747 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"06e262448da43be1f59f8206d3ca149b097157a5612b8e5d368736aac9a62a57"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170795 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"ba8102ea51bc30c759201a866a8431d1baaf2f6f474cc96b36529ea5a04451ec"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170805 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"cd237a014093f689d7df188672e570ebe94fb2b0f49e1d1a6c58473d64c3b740"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170815 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"b18142ab07c004a98bff59b529be5e4c49e5ff54caf77526350e39ce8c20c39c"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170823 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"cd51894ecb6651dbb869d73d13d130a6f2fb7d28b0f7980f406450f51c4449b9"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170831 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"95b99cc19f9b362efc74b2612128d2a9a4247e4bfa40d9b7ca5e1c764c78c5c9"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170871 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"5fb020c3045a74d5048d67e9e8f939b0d2f9348bab133b9b65f5cfb8ff321504"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170879 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"22ba6a3f7d0c659cde8f9b4b899df8e9b5b426c7a3f7e76b86ffc508196c5584"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.170911 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"a461381cdfc3c2eee74eb1bdcf7d5ebfdb0fb05a83a09c83b9bcd518fb51b2d5"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.179642 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.179679 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.182070 4935 scope.go:117] "RemoveContainer" containerID="9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.182927 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f27b0edc-d2a4-418e-8c62-82f1287cc1b4" (UID: "f27b0edc-d2a4-418e-8c62-82f1287cc1b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.187960 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.188209 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="f950314f-27aa-4a44-b13a-4b4f3a7495ab" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.193730 4935 generic.go:334] "Generic (PLEG): container finished" podID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerID="8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a" exitCode=143 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.193991 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" event={"ID":"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b","Type":"ContainerDied","Data":"8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.201581 4935 generic.go:334] "Generic (PLEG): container finished" podID="aac8194b-9784-4eed-8c35-fb65a584b525" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.201662 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pl4nq" event={"ID":"aac8194b-9784-4eed-8c35-fb65a584b525","Type":"ContainerDied","Data":"81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.210381 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.214074 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe76393-92f2-4164-a549-5f75ced24fb3-kube-api-access-8rtjh" (OuterVolumeSpecName: "kube-api-access-8rtjh") pod "abe76393-92f2-4164-a549-5f75ced24fb3" (UID: "abe76393-92f2-4164-a549-5f75ced24fb3"). InnerVolumeSpecName "kube-api-access-8rtjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.215456 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.224311 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.228298 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" containerName="nova-scheduler-scheduler" containerID="cri-o://2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.228515 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="77853c76-2bbe-42e5-a0ef-03d5c5fe2402" containerName="nova-cell1-conductor-conductor" containerID="cri-o://f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.241663 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vlqf6"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.282207 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.282242 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rtjh\" (UniqueName: \"kubernetes.io/projected/abe76393-92f2-4164-a549-5f75ced24fb3-kube-api-access-8rtjh\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.309664 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vlqf6"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.320702 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerName="rabbitmq" containerID="cri-o://9e3ecf50ee2e44f330efce8b076a27f4149d1c011a741f1c565a07c5137e9563" gracePeriod=604800 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.326960 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxztf"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.327755 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d0d984f8-20c4-4b30-b767-c6106aa279cf/ovsdbserver-sb/0.log" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.327902 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0d984f8-20c4-4b30-b767-c6106aa279cf","Type":"ContainerDied","Data":"8021a795afeb72e1c25c850f463e20c2b39ebf74611e8d459eb501f0960c8e2d"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.328100 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.352955 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.353432 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" containerName="nova-cell0-conductor-conductor" containerID="cri-o://24a61e5d809f2202e8aff2ffc919b4f809f3ea3e05132213aaafa01f0e1fc0d4" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.361346 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wxztf"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.398075 4935 scope.go:117] "RemoveContainer" containerID="d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f" Oct 05 07:15:51 crc kubenswrapper[4935]: E1005 07:15:51.407734 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f\": container with ID starting with d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f not found: ID does not exist" containerID="d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.407772 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f"} err="failed to get container status \"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f\": rpc error: code = NotFound desc = could not find container \"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f\": container with ID starting with d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f not found: ID does not exist" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.407796 4935 scope.go:117] "RemoveContainer" containerID="9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.407869 4935 generic.go:334] "Generic (PLEG): container finished" podID="27a28306-70be-4556-9659-64999f775195" containerID="16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000" exitCode=0 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.408215 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56bb7fd5-dmzld" event={"ID":"27a28306-70be-4556-9659-64999f775195","Type":"ContainerDied","Data":"16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.411344 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "d0d984f8-20c4-4b30-b767-c6106aa279cf" (UID: "d0d984f8-20c4-4b30-b767-c6106aa279cf"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: E1005 07:15:51.421076 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95\": container with ID starting with 9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95 not found: ID does not exist" containerID="9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.421256 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95"} err="failed to get container status \"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95\": rpc error: code = NotFound desc = could not find container \"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95\": container with ID starting with 9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95 not found: ID does not exist" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.421333 4935 scope.go:117] "RemoveContainer" containerID="d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.422164 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f"} err="failed to get container status \"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f\": rpc error: code = NotFound desc = could not find container \"d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f\": container with ID starting with d6161b5a3624fd7cdd1675966af2e6b69a9d686b51df19ffa64b08cfb508893f not found: ID does not exist" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.422203 4935 scope.go:117] "RemoveContainer" containerID="9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.423701 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95"} err="failed to get container status \"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95\": rpc error: code = NotFound desc = could not find container \"9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95\": container with ID starting with 9ec7b8c0b1909b05d1066ae63847f23db25f550f503aff1e352259c3d9f78f95 not found: ID does not exist" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.423728 4935 scope.go:117] "RemoveContainer" containerID="453dce919ffa6f7b445b83173fc85bcadcbbe96d524dbe18eb42e012aa8b54c1" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.426227 4935 generic.go:334] "Generic (PLEG): container finished" podID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerID="9b2b88228c11fcf40bd73c44aadf5f4a93619dfc6b49a5c328edcf98f4d93a7e" exitCode=143 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.426301 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74989f54d8-bw287" event={"ID":"3c422d4a-a835-443a-9c6d-3161bc80c1aa","Type":"ContainerDied","Data":"9b2b88228c11fcf40bd73c44aadf5f4a93619dfc6b49a5c328edcf98f4d93a7e"} Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.436171 4935 generic.go:334] "Generic (PLEG): container finished" podID="c9b2e8af-97b0-4103-a241-ca473f299043" containerID="a5103fe4899d57404d97b7ee8251771675c6a622925145cac9caa880ee1b5c91" exitCode=143 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.436217 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57fd77d677-fr4zj" event={"ID":"c9b2e8af-97b0-4103-a241-ca473f299043","Type":"ContainerDied","Data":"a5103fe4899d57404d97b7ee8251771675c6a622925145cac9caa880ee1b5c91"} Oct 05 07:15:51 crc kubenswrapper[4935]: E1005 07:15:51.437469 4935 kuberuntime_gc.go:389] "Failed to remove container log dead symlink" err="remove /var/log/containers/ovsdbserver-sb-0_openstack_openstack-network-exporter-453dce919ffa6f7b445b83173fc85bcadcbbe96d524dbe18eb42e012aa8b54c1.log: no such file or directory" path="/var/log/containers/ovsdbserver-sb-0_openstack_openstack-network-exporter-453dce919ffa6f7b445b83173fc85bcadcbbe96d524dbe18eb42e012aa8b54c1.log" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.452577 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abe76393-92f2-4164-a549-5f75ced24fb3" (UID: "abe76393-92f2-4164-a549-5f75ced24fb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: W1005 07:15:51.462676 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6ccd5ed_23d9_4d0b_99fb_68655c0199b5.slice/crio-5d029f67b11fb02bbf243cb6284984243ab10f24e192066df644eec5181c70fa WatchSource:0}: Error finding container 5d029f67b11fb02bbf243cb6284984243ab10f24e192066df644eec5181c70fa: Status 404 returned error can't find the container with id 5d029f67b11fb02bbf243cb6284984243ab10f24e192066df644eec5181c70fa Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.462726 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6f6d895595-jv76b"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.462972 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6f6d895595-jv76b" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerName="proxy-httpd" containerID="cri-o://60a3d290ffb9e90b0086750d7b0d89ef70348c9efbf1a8dea1b64dbf0d73f9de" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.463401 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6f6d895595-jv76b" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerName="proxy-server" containerID="cri-o://20a20a4e69bdd787df49c91056089308888f8504fc31300ec9ca36a389aeb510" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.475663 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-l5t6l"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.479401 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "f27b0edc-d2a4-418e-8c62-82f1287cc1b4" (UID: "f27b0edc-d2a4-418e-8c62-82f1287cc1b4"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.487116 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-l5t6l"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.490208 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb3bb-account-delete-v2qzp"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.491795 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d984f8-20c4-4b30-b767-c6106aa279cf-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.491815 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.491825 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.510936 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.530357 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "f27b0edc-d2a4-418e-8c62-82f1287cc1b4" (UID: "f27b0edc-d2a4-418e-8c62-82f1287cc1b4"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.544491 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "abe76393-92f2-4164-a549-5f75ced24fb3" (UID: "abe76393-92f2-4164-a549-5f75ced24fb3"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.593961 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27b0edc-d2a4-418e-8c62-82f1287cc1b4-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.594195 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.594205 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.603060 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "abe76393-92f2-4164-a549-5f75ced24fb3" (UID: "abe76393-92f2-4164-a549-5f75ced24fb3"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.603227 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="7bedc55a-2396-4249-ac54-a753a67dca72" containerName="galera" containerID="cri-o://5985a42517f3f6b261d93f0ab627c53c5bc297b9e7699a332b44b1e66d30ab31" gracePeriod=30 Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.672682 4935 scope.go:117] "RemoveContainer" containerID="d3e5a40a1e26ef3caaba5145609516f5ff29e761ac28f571d0da1f626e989083" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.697051 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/abe76393-92f2-4164-a549-5f75ced24fb3-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.747833 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.780619 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.798392 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.799550 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-svc\") pod \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.799662 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-nb\") pod \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.799683 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4fxg\" (UniqueName: \"kubernetes.io/projected/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-kube-api-access-b4fxg\") pod \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.799734 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-config\") pod \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.799824 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-swift-storage-0\") pod \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.799945 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-sb\") pod \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\" (UID: \"8bf67c05-04fb-414a-aa6e-8cdc3831cb94\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.809006 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementdd1f-account-delete-cgh8g" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.813184 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-kube-api-access-b4fxg" (OuterVolumeSpecName: "kube-api-access-b4fxg") pod "8bf67c05-04fb-414a-aa6e-8cdc3831cb94" (UID: "8bf67c05-04fb-414a-aa6e-8cdc3831cb94"). InnerVolumeSpecName "kube-api-access-b4fxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.840241 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.846112 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.903735 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znz9r\" (UniqueName: \"kubernetes.io/projected/0095740d-765a-4cba-b0c8-d36592fad800-kube-api-access-znz9r\") pod \"0095740d-765a-4cba-b0c8-d36592fad800\" (UID: \"0095740d-765a-4cba-b0c8-d36592fad800\") " Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.904192 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4fxg\" (UniqueName: \"kubernetes.io/projected/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-kube-api-access-b4fxg\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.927047 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0095740d-765a-4cba-b0c8-d36592fad800-kube-api-access-znz9r" (OuterVolumeSpecName: "kube-api-access-znz9r") pod "0095740d-765a-4cba-b0c8-d36592fad800" (UID: "0095740d-765a-4cba-b0c8-d36592fad800"). InnerVolumeSpecName "kube-api-access-znz9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.985224 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8bf67c05-04fb-414a-aa6e-8cdc3831cb94" (UID: "8bf67c05-04fb-414a-aa6e-8cdc3831cb94"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:51 crc kubenswrapper[4935]: I1005 07:15:51.994263 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-config" (OuterVolumeSpecName: "config") pod "8bf67c05-04fb-414a-aa6e-8cdc3831cb94" (UID: "8bf67c05-04fb-414a-aa6e-8cdc3831cb94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.004165 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi70bb-account-delete-5wwbr"] Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.007933 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.007961 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.007973 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znz9r\" (UniqueName: \"kubernetes.io/projected/0095740d-765a-4cba-b0c8-d36592fad800-kube-api-access-znz9r\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.011941 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8bf67c05-04fb-414a-aa6e-8cdc3831cb94" (UID: "8bf67c05-04fb-414a-aa6e-8cdc3831cb94"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.031446 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanb178-account-delete-rnqtn"] Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.033835 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell09d1d-account-delete-24dn6"] Oct 05 07:15:52 crc kubenswrapper[4935]: W1005 07:15:52.060274 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod819c9cf6_f5e5_46af_9fee_c1c4cc88c29a.slice/crio-ea0f16f5c07f8977895c2b0ae19b218ccee8c33ccdd453ed2659ca7da42027fa WatchSource:0}: Error finding container ea0f16f5c07f8977895c2b0ae19b218ccee8c33ccdd453ed2659ca7da42027fa: Status 404 returned error can't find the container with id ea0f16f5c07f8977895c2b0ae19b218ccee8c33ccdd453ed2659ca7da42027fa Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.070644 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder9bb0-account-delete-8m44f"] Oct 05 07:15:52 crc kubenswrapper[4935]: W1005 07:15:52.073521 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7143940e_eedf_48a0_b0dc_4ec91df08ad0.slice/crio-292cd469a0e080e6ac3d0a3525828e356bb30257f2ad94878d5ea283b0b91232 WatchSource:0}: Error finding container 292cd469a0e080e6ac3d0a3525828e356bb30257f2ad94878d5ea283b0b91232: Status 404 returned error can't find the container with id 292cd469a0e080e6ac3d0a3525828e356bb30257f2ad94878d5ea283b0b91232 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.073837 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8bf67c05-04fb-414a-aa6e-8cdc3831cb94" (UID: "8bf67c05-04fb-414a-aa6e-8cdc3831cb94"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.076623 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancee563-account-delete-dqlrz"] Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.089316 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8bf67c05-04fb-414a-aa6e-8cdc3831cb94" (UID: "8bf67c05-04fb-414a-aa6e-8cdc3831cb94"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.109580 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.109606 4935 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.109616 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf67c05-04fb-414a-aa6e-8cdc3831cb94-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.109679 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.109725 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data podName:8dcf22e2-2acd-4b49-a2bc-813a50f89409 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:56.109708297 +0000 UTC m=+1389.992334757 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data") pod "rabbitmq-server-0" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409") : configmap "rabbitmq-config-data" not found Oct 05 07:15:52 crc kubenswrapper[4935]: W1005 07:15:52.140788 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod420a72a6_506a_42e6_b77a_ee5ca8527768.slice/crio-5f03f2be56a1ab2560e59e271a14c16c08d5f954a29cb8888f9f85299e881a6e WatchSource:0}: Error finding container 5f03f2be56a1ab2560e59e271a14c16c08d5f954a29cb8888f9f85299e881a6e: Status 404 returned error can't find the container with id 5f03f2be56a1ab2560e59e271a14c16c08d5f954a29cb8888f9f85299e881a6e Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.260135 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.263788 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:15:52 crc kubenswrapper[4935]: W1005 07:15:52.264181 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bd6e7de_59a5_479d_b2b8_7f060883f6c8.slice/crio-6419b79b1d539c5af236dd70c7f6b0764bf7a78b49dfc9d1ca0e2802cc4bae5e WatchSource:0}: Error finding container 6419b79b1d539c5af236dd70c7f6b0764bf7a78b49dfc9d1ca0e2802cc4bae5e: Status 404 returned error can't find the container with id 6419b79b1d539c5af236dd70c7f6b0764bf7a78b49dfc9d1ca0e2802cc4bae5e Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.290752 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.299487 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.299576 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.299984 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.303793 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.303878 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.304075 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.421402 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx7mb\" (UniqueName: \"kubernetes.io/projected/f950314f-27aa-4a44-b13a-4b4f3a7495ab-kube-api-access-zx7mb\") pod \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.421508 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-nova-novncproxy-tls-certs\") pod \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.421571 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-vencrypt-tls-certs\") pod \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.421614 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-config-data\") pod \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.421749 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-combined-ca-bundle\") pod \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\" (UID: \"f950314f-27aa-4a44-b13a-4b4f3a7495ab\") " Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.456134 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f950314f-27aa-4a44-b13a-4b4f3a7495ab-kube-api-access-zx7mb" (OuterVolumeSpecName: "kube-api-access-zx7mb") pod "f950314f-27aa-4a44-b13a-4b4f3a7495ab" (UID: "f950314f-27aa-4a44-b13a-4b4f3a7495ab"). InnerVolumeSpecName "kube-api-access-zx7mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.472636 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-config-data" (OuterVolumeSpecName: "config-data") pod "f950314f-27aa-4a44-b13a-4b4f3a7495ab" (UID: "f950314f-27aa-4a44-b13a-4b4f3a7495ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.476051 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" event={"ID":"8bf67c05-04fb-414a-aa6e-8cdc3831cb94","Type":"ContainerDied","Data":"114eb74ef5792ce6ed05052372361ebe58fe546bc2e8963abcc523b948677238"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.476107 4935 scope.go:117] "RemoveContainer" containerID="5c401639335e92358e96e3ee9d7c9cbcdd83e9b6aa688f56a6c57891805ffe6d" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.476248 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.496647 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder9bb0-account-delete-8m44f" event={"ID":"420a72a6-506a-42e6-b77a-ee5ca8527768","Type":"ContainerStarted","Data":"5f03f2be56a1ab2560e59e271a14c16c08d5f954a29cb8888f9f85299e881a6e"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.499005 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.504323 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "f950314f-27aa-4a44-b13a-4b4f3a7495ab" (UID: "f950314f-27aa-4a44-b13a-4b4f3a7495ab"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.505085 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f950314f-27aa-4a44-b13a-4b4f3a7495ab" (UID: "f950314f-27aa-4a44-b13a-4b4f3a7495ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.505178 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancee563-account-delete-dqlrz" event={"ID":"6bd6e7de-59a5-479d-b2b8-7f060883f6c8","Type":"ContainerStarted","Data":"6419b79b1d539c5af236dd70c7f6b0764bf7a78b49dfc9d1ca0e2802cc4bae5e"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.506330 4935 generic.go:334] "Generic (PLEG): container finished" podID="d6ccd5ed-23d9-4d0b-99fb-68655c0199b5" containerID="2d66e9ffc7b8a645d646583394f15aa468c427b2fc5f36b059ac5211365b9e21" exitCode=0 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.506370 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb3bb-account-delete-v2qzp" event={"ID":"d6ccd5ed-23d9-4d0b-99fb-68655c0199b5","Type":"ContainerDied","Data":"2d66e9ffc7b8a645d646583394f15aa468c427b2fc5f36b059ac5211365b9e21"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.506386 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb3bb-account-delete-v2qzp" event={"ID":"d6ccd5ed-23d9-4d0b-99fb-68655c0199b5","Type":"ContainerStarted","Data":"5d029f67b11fb02bbf243cb6284984243ab10f24e192066df644eec5181c70fa"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.523983 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.524112 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx7mb\" (UniqueName: \"kubernetes.io/projected/f950314f-27aa-4a44-b13a-4b4f3a7495ab-kube-api-access-zx7mb\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.524169 4935 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.524319 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.524366 4935 secret.go:188] Couldn't get secret openstack/placement-scripts: secret "placement-scripts" not found Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.524543 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts podName:88a7ceab-efbb-4ba4-a5f2-c6702abbc314 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:56.524528342 +0000 UTC m=+1390.407154802 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts") pod "placement-6dbdfd7554-jrf6q" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314") : secret "placement-scripts" not found Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.524394 4935 secret.go:188] Couldn't get secret openstack/placement-config-data: secret "placement-config-data" not found Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.524696 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data podName:88a7ceab-efbb-4ba4-a5f2-c6702abbc314 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:56.524688756 +0000 UTC m=+1390.407315216 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data") pod "placement-6dbdfd7554-jrf6q" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314") : secret "placement-config-data" not found Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.526394 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "f950314f-27aa-4a44-b13a-4b4f3a7495ab" (UID: "f950314f-27aa-4a44-b13a-4b4f3a7495ab"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.528546 4935 generic.go:334] "Generic (PLEG): container finished" podID="f950314f-27aa-4a44-b13a-4b4f3a7495ab" containerID="1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd" exitCode=0 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.528880 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f950314f-27aa-4a44-b13a-4b4f3a7495ab","Type":"ContainerDied","Data":"1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.529003 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f950314f-27aa-4a44-b13a-4b4f3a7495ab","Type":"ContainerDied","Data":"b61e8ab4c7ba520380864b515362aca5c445949eca3b1e4d2053f94929164407"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.529199 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.544300 4935 generic.go:334] "Generic (PLEG): container finished" podID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerID="20a20a4e69bdd787df49c91056089308888f8504fc31300ec9ca36a389aeb510" exitCode=0 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.544350 4935 generic.go:334] "Generic (PLEG): container finished" podID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerID="60a3d290ffb9e90b0086750d7b0d89ef70348c9efbf1a8dea1b64dbf0d73f9de" exitCode=0 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.544397 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f6d895595-jv76b" event={"ID":"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba","Type":"ContainerDied","Data":"20a20a4e69bdd787df49c91056089308888f8504fc31300ec9ca36a389aeb510"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.544457 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f6d895595-jv76b" event={"ID":"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba","Type":"ContainerDied","Data":"60a3d290ffb9e90b0086750d7b0d89ef70348c9efbf1a8dea1b64dbf0d73f9de"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.546516 4935 generic.go:334] "Generic (PLEG): container finished" podID="77853c76-2bbe-42e5-a0ef-03d5c5fe2402" containerID="f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465" exitCode=0 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.546594 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"77853c76-2bbe-42e5-a0ef-03d5c5fe2402","Type":"ContainerDied","Data":"f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.546632 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"77853c76-2bbe-42e5-a0ef-03d5c5fe2402","Type":"ContainerDied","Data":"cd8f2c98f54599f7e46b0b16cd909e8d1b6113aa2db919860d5ca310abd5790f"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.546651 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd8f2c98f54599f7e46b0b16cd909e8d1b6113aa2db919860d5ca310abd5790f" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.548077 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementdd1f-account-delete-cgh8g" event={"ID":"0095740d-765a-4cba-b0c8-d36592fad800","Type":"ContainerDied","Data":"30792e5f0d29f0148506ee6aca3f68261e7c621815926d30627de0c32c6af5f9"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.548147 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementdd1f-account-delete-cgh8g" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.576416 4935 generic.go:334] "Generic (PLEG): container finished" podID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerID="3e0b1900e020511c33b937275937ce79dc33f680831c304611354b614ac0d6a6" exitCode=143 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.576487 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c","Type":"ContainerDied","Data":"3e0b1900e020511c33b937275937ce79dc33f680831c304611354b614ac0d6a6"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.579455 4935 generic.go:334] "Generic (PLEG): container finished" podID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerID="7378ebd5875058e6928f9504c6ce673a929f0b593e3c881ba94728592aaa6422" exitCode=0 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.579599 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7c7acc7a-797f-46cd-b298-1f28fd2951d6","Type":"ContainerDied","Data":"7378ebd5875058e6928f9504c6ce673a929f0b593e3c881ba94728592aaa6422"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.579694 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7c7acc7a-797f-46cd-b298-1f28fd2951d6","Type":"ContainerDied","Data":"7d70d45c48415ff1f6889a218122105e1192896f2ffe16de0dff780b19f7a4c5"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.579781 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d70d45c48415ff1f6889a218122105e1192896f2ffe16de0dff780b19f7a4c5" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.584857 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanb178-account-delete-rnqtn" event={"ID":"b732474b-e717-4d8d-91b7-e00409800f85","Type":"ContainerStarted","Data":"ebe205a0654a7d61ca4c49ac56972ed1cdc7d59980a0fcfd9437cd3e40829de1"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.587028 4935 generic.go:334] "Generic (PLEG): container finished" podID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerID="122b7e3178e70d3a5f9d4ca95b47f2e8787d8b3c771046206b29a4dc71f41a66" exitCode=143 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.587168 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f1eb003-e111-4a86-b6c7-a899aa0426a0","Type":"ContainerDied","Data":"122b7e3178e70d3a5f9d4ca95b47f2e8787d8b3c771046206b29a4dc71f41a66"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.597143 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi70bb-account-delete-5wwbr" event={"ID":"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a","Type":"ContainerStarted","Data":"ea0f16f5c07f8977895c2b0ae19b218ccee8c33ccdd453ed2659ca7da42027fa"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.598154 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell09d1d-account-delete-24dn6" event={"ID":"7143940e-eedf-48a0-b0dc-4ec91df08ad0","Type":"ContainerStarted","Data":"292cd469a0e080e6ac3d0a3525828e356bb30257f2ad94878d5ea283b0b91232"} Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.626109 4935 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f950314f-27aa-4a44-b13a-4b4f3a7495ab-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.682767 4935 scope.go:117] "RemoveContainer" containerID="90e177c52c5d7180cd73793a31e9ceb0eeac1791a8bf7d225dc5c4a85800b727" Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.694605 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465 is running failed: container process not found" containerID="f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.695043 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465 is running failed: container process not found" containerID="f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.695385 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465 is running failed: container process not found" containerID="f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:15:52 crc kubenswrapper[4935]: E1005 07:15:52.695489 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="77853c76-2bbe-42e5-a0ef-03d5c5fe2402" containerName="nova-cell1-conductor-conductor" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.789234 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="234e0c62-014f-40c5-834f-e897b9593cdb" path="/var/lib/kubelet/pods/234e0c62-014f-40c5-834f-e897b9593cdb/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.789784 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="319fa1f0-1aeb-4881-a883-4b2f7b61f833" path="/var/lib/kubelet/pods/319fa1f0-1aeb-4881-a883-4b2f7b61f833/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.790532 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61acbc8c-29cc-4afa-b3ce-34d65be5a32e" path="/var/lib/kubelet/pods/61acbc8c-29cc-4afa-b3ce-34d65be5a32e/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.791121 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65623cb0-f66d-4bfd-b732-680ad4da5c1f" path="/var/lib/kubelet/pods/65623cb0-f66d-4bfd-b732-680ad4da5c1f/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.792274 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94914bdc-6fed-4035-a624-93feacf3ba00" path="/var/lib/kubelet/pods/94914bdc-6fed-4035-a624-93feacf3ba00/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.793076 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d26dccc-7629-42a9-9597-600f4750c11b" path="/var/lib/kubelet/pods/9d26dccc-7629-42a9-9597-600f4750c11b/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.793687 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c2d664-55ad-4316-849a-b05d5a439388" path="/var/lib/kubelet/pods/a1c2d664-55ad-4316-849a-b05d5a439388/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.794585 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3f78262-5eb9-4a4d-9a93-2333f428b032" path="/var/lib/kubelet/pods/a3f78262-5eb9-4a4d-9a93-2333f428b032/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.796527 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abe76393-92f2-4164-a549-5f75ced24fb3" path="/var/lib/kubelet/pods/abe76393-92f2-4164-a549-5f75ced24fb3/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.799515 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca591a2f-54ed-41b1-aa10-be879b218844" path="/var/lib/kubelet/pods/ca591a2f-54ed-41b1-aa10-be879b218844/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.801030 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" path="/var/lib/kubelet/pods/d0d984f8-20c4-4b30-b767-c6106aa279cf/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.801851 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" path="/var/lib/kubelet/pods/f27b0edc-d2a4-418e-8c62-82f1287cc1b4/volumes" Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.873562 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.874248 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="ceilometer-central-agent" containerID="cri-o://1a7bd0c309de5a9b869370ed377d8a46c8a2f0c566fa6724d9ed2c87fbfb0e4a" gracePeriod=30 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.874797 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="proxy-httpd" containerID="cri-o://fc4511ccd783fd2b8c6cb4dce5c4f3cc88512854346a9a9f11bfe9081fb534e4" gracePeriod=30 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.874869 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="sg-core" containerID="cri-o://48ae75168e8a69d6f8a97da4dacb4141e0980a9ae4a1946601d92425b5f4f413" gracePeriod=30 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.874942 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="ceilometer-notification-agent" containerID="cri-o://37c00f2d3e9897fa787e5b2b4870c3b4e33ad0c59b0775763a7aeb443b6567b7" gracePeriod=30 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.898371 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.898628 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c710b93b-c2fe-4a03-ba8c-f280a9d67da6" containerName="kube-state-metrics" containerID="cri-o://da3d1180f3ad1963dd2711515fe79e75473abc8f27f2f472d500b4a81d01765f" gracePeriod=30 Oct 05 07:15:52 crc kubenswrapper[4935]: I1005 07:15:52.940124 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.018879 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.019471 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="b82213d6-93fa-48fc-808b-36f902cce15c" containerName="memcached" containerID="cri-o://fb6f5493ca2adedadd805d4c56c29f8cf8470abc87c240d1db6707f07565dd60" gracePeriod=30 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.064624 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c7acc7a-797f-46cd-b298-1f28fd2951d6-etc-machine-id\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.064663 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjx6t\" (UniqueName: \"kubernetes.io/projected/7c7acc7a-797f-46cd-b298-1f28fd2951d6-kube-api-access-jjx6t\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.064736 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data-custom\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.064825 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-internal-tls-certs\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.064848 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-combined-ca-bundle\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.064940 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-scripts\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.064996 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.065012 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c7acc7a-797f-46cd-b298-1f28fd2951d6-logs\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.065024 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c7acc7a-797f-46cd-b298-1f28fd2951d6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.065043 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-public-tls-certs\") pod \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\" (UID: \"7c7acc7a-797f-46cd-b298-1f28fd2951d6\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.065950 4935 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c7acc7a-797f-46cd-b298-1f28fd2951d6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.087812 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c7acc7a-797f-46cd-b298-1f28fd2951d6-kube-api-access-jjx6t" (OuterVolumeSpecName: "kube-api-access-jjx6t") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "kube-api-access-jjx6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.094819 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-scripts" (OuterVolumeSpecName: "scripts") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.117009 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c7acc7a-797f-46cd-b298-1f28fd2951d6-logs" (OuterVolumeSpecName: "logs") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.131815 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.138921 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-7s8ss"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.147883 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-79d9bc6587-tvhv2"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.148184 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-79d9bc6587-tvhv2" podUID="6214d810-93b4-4537-a655-642df573048b" containerName="keystone-api" containerID="cri-o://010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413" gracePeriod=30 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.156920 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data" (OuterVolumeSpecName: "config-data") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.165177 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.173354 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.173380 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.173389 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.173399 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c7acc7a-797f-46cd-b298-1f28fd2951d6-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.173408 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjx6t\" (UniqueName: \"kubernetes.io/projected/7c7acc7a-797f-46cd-b298-1f28fd2951d6-kube-api-access-jjx6t\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.173416 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: E1005 07:15:53.173479 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:53 crc kubenswrapper[4935]: E1005 07:15:53.173526 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data podName:53fe24e1-5efb-4422-ade0-16e315b7e328 nodeName:}" failed. No retries permitted until 2025-10-05 07:15:57.173510383 +0000 UTC m=+1391.056136843 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data") pod "rabbitmq-cell1-server-0" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.173672 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-7s8ss"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.191693 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-k2ff8"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.197509 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-k2ff8"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.202453 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.208797 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-dhmhx"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.213867 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.216034 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-dhmhx"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.221486 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5ce4-account-create-n7jk5"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.226133 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5ce4-account-create-n7jk5"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.228092 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7c7acc7a-797f-46cd-b298-1f28fd2951d6" (UID: "7c7acc7a-797f-46cd-b298-1f28fd2951d6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.274688 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.274707 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c7acc7a-797f-46cd-b298-1f28fd2951d6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.294040 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.197:3000/\": dial tcp 10.217.0.197:3000: connect: connection reset by peer" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.409333 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" containerName="galera" containerID="cri-o://09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d" gracePeriod=30 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.612652 4935 scope.go:117] "RemoveContainer" containerID="1d9ad091dc634c8651d67b5d664e7fdae2e562baf23016a53943168af716e873" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.613269 4935 generic.go:334] "Generic (PLEG): container finished" podID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerID="fc4511ccd783fd2b8c6cb4dce5c4f3cc88512854346a9a9f11bfe9081fb534e4" exitCode=0 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.613305 4935 generic.go:334] "Generic (PLEG): container finished" podID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerID="48ae75168e8a69d6f8a97da4dacb4141e0980a9ae4a1946601d92425b5f4f413" exitCode=2 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.613318 4935 generic.go:334] "Generic (PLEG): container finished" podID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerID="1a7bd0c309de5a9b869370ed377d8a46c8a2f0c566fa6724d9ed2c87fbfb0e4a" exitCode=0 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.613366 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerDied","Data":"fc4511ccd783fd2b8c6cb4dce5c4f3cc88512854346a9a9f11bfe9081fb534e4"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.613396 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerDied","Data":"48ae75168e8a69d6f8a97da4dacb4141e0980a9ae4a1946601d92425b5f4f413"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.613409 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerDied","Data":"1a7bd0c309de5a9b869370ed377d8a46c8a2f0c566fa6724d9ed2c87fbfb0e4a"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.624541 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb3bb-account-delete-v2qzp" event={"ID":"d6ccd5ed-23d9-4d0b-99fb-68655c0199b5","Type":"ContainerDied","Data":"5d029f67b11fb02bbf243cb6284984243ab10f24e192066df644eec5181c70fa"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.624605 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d029f67b11fb02bbf243cb6284984243ab10f24e192066df644eec5181c70fa" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.628178 4935 generic.go:334] "Generic (PLEG): container finished" podID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerID="24b5a0ee97d3993b21bd1da36e2ba46a94d38c77257bf5f70b85ea35748a294c" exitCode=0 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.628267 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9acce3fe-5a42-49e7-b3e7-a72f01af5736","Type":"ContainerDied","Data":"24b5a0ee97d3993b21bd1da36e2ba46a94d38c77257bf5f70b85ea35748a294c"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.628298 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9acce3fe-5a42-49e7-b3e7-a72f01af5736","Type":"ContainerDied","Data":"d5a662fd065201895283c03ffbb32808389eddd8ab036e44f1677d1a8f7b9801"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.628310 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5a662fd065201895283c03ffbb32808389eddd8ab036e44f1677d1a8f7b9801" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.630247 4935 generic.go:334] "Generic (PLEG): container finished" podID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerID="829086067b4c67ac2011dcae6d0f8b4b1d850acfb49e4cf8b466a30f81bb3be2" exitCode=0 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.630310 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbdfd7554-jrf6q" event={"ID":"88a7ceab-efbb-4ba4-a5f2-c6702abbc314","Type":"ContainerDied","Data":"829086067b4c67ac2011dcae6d0f8b4b1d850acfb49e4cf8b466a30f81bb3be2"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.635469 4935 generic.go:334] "Generic (PLEG): container finished" podID="7bedc55a-2396-4249-ac54-a753a67dca72" containerID="5985a42517f3f6b261d93f0ab627c53c5bc297b9e7699a332b44b1e66d30ab31" exitCode=0 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.635540 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7bedc55a-2396-4249-ac54-a753a67dca72","Type":"ContainerDied","Data":"5985a42517f3f6b261d93f0ab627c53c5bc297b9e7699a332b44b1e66d30ab31"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.635586 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7bedc55a-2396-4249-ac54-a753a67dca72","Type":"ContainerDied","Data":"857b569be287a3227232d02e565e22a5e29684e717b08f7787146b2c47045af3"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.635601 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="857b569be287a3227232d02e565e22a5e29684e717b08f7787146b2c47045af3" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.643290 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6f6d895595-jv76b" event={"ID":"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba","Type":"ContainerDied","Data":"65e13bd44278107300790a93bc76ecae382f285e1385cee5ef8b18b0a1bbed0b"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.643334 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65e13bd44278107300790a93bc76ecae382f285e1385cee5ef8b18b0a1bbed0b" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.646322 4935 generic.go:334] "Generic (PLEG): container finished" podID="c710b93b-c2fe-4a03-ba8c-f280a9d67da6" containerID="da3d1180f3ad1963dd2711515fe79e75473abc8f27f2f472d500b4a81d01765f" exitCode=2 Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.646381 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c710b93b-c2fe-4a03-ba8c-f280a9d67da6","Type":"ContainerDied","Data":"da3d1180f3ad1963dd2711515fe79e75473abc8f27f2f472d500b4a81d01765f"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.646413 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c710b93b-c2fe-4a03-ba8c-f280a9d67da6","Type":"ContainerDied","Data":"e3f24b7f7c6f483cb8122e57d065e46e001d54a174b2fb043d4bfb249336b0ee"} Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.646421 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.646427 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3f24b7f7c6f483cb8122e57d065e46e001d54a174b2fb043d4bfb249336b0ee" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.696679 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.718655 4935 scope.go:117] "RemoveContainer" containerID="1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.782082 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-combined-ca-bundle\") pod \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.782127 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-config-data\") pod \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.782194 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcdqm\" (UniqueName: \"kubernetes.io/projected/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-kube-api-access-pcdqm\") pod \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\" (UID: \"77853c76-2bbe-42e5-a0ef-03d5c5fe2402\") " Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.803321 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-kube-api-access-pcdqm" (OuterVolumeSpecName: "kube-api-access-pcdqm") pod "77853c76-2bbe-42e5-a0ef-03d5c5fe2402" (UID: "77853c76-2bbe-42e5-a0ef-03d5c5fe2402"). InnerVolumeSpecName "kube-api-access-pcdqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.819328 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77853c76-2bbe-42e5-a0ef-03d5c5fe2402" (UID: "77853c76-2bbe-42e5-a0ef-03d5c5fe2402"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.826241 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-config-data" (OuterVolumeSpecName: "config-data") pod "77853c76-2bbe-42e5-a0ef-03d5c5fe2402" (UID: "77853c76-2bbe-42e5-a0ef-03d5c5fe2402"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.885157 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcdqm\" (UniqueName: \"kubernetes.io/projected/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-kube-api-access-pcdqm\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.885189 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.885198 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77853c76-2bbe-42e5-a0ef-03d5c5fe2402-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.934315 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-wfm9k"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.952213 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-wfm9k"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.958558 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9d1d-account-create-64df8"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.964674 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell09d1d-account-delete-24dn6"] Oct 05 07:15:53 crc kubenswrapper[4935]: I1005 07:15:53.970954 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9d1d-account-create-64df8"] Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.021260 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb3bb-account-delete-v2qzp" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.069784 4935 scope.go:117] "RemoveContainer" containerID="1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd" Oct 05 07:15:54 crc kubenswrapper[4935]: E1005 07:15:54.083156 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd\": container with ID starting with 1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd not found: ID does not exist" containerID="1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.083379 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd"} err="failed to get container status \"1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd\": rpc error: code = NotFound desc = could not find container \"1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd\": container with ID starting with 1d3f70c4b8b0ab09a359345e4b9a29821121adb3bb94fe9fe1eb88ec67b4d1bd not found: ID does not exist" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.084013 4935 scope.go:117] "RemoveContainer" containerID="d2d2b4c5eb14625e816f7c37ac0375d97c1c1a9134db1e83ded9f96ed1398e3b" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.088645 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7fqw\" (UniqueName: \"kubernetes.io/projected/d6ccd5ed-23d9-4d0b-99fb-68655c0199b5-kube-api-access-l7fqw\") pod \"d6ccd5ed-23d9-4d0b-99fb-68655c0199b5\" (UID: \"d6ccd5ed-23d9-4d0b-99fb-68655c0199b5\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.094034 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6ccd5ed-23d9-4d0b-99fb-68655c0199b5-kube-api-access-l7fqw" (OuterVolumeSpecName: "kube-api-access-l7fqw") pod "d6ccd5ed-23d9-4d0b-99fb-68655c0199b5" (UID: "d6ccd5ed-23d9-4d0b-99fb-68655c0199b5"). InnerVolumeSpecName "kube-api-access-l7fqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.107053 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.123291 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.143930 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.154845 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.163697 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.172285 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": read tcp 10.217.0.2:55988->10.217.0.203:8775: read: connection reset by peer" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.172617 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": read tcp 10.217.0.2:55980->10.217.0.203:8775: read: connection reset by peer" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.175280 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194109 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-galera-tls-certs\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194152 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-log-httpd\") pod \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194171 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-internal-tls-certs\") pod \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194216 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-combined-ca-bundle\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194242 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-operator-scripts\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194269 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-config-data\") pod \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194293 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5mws\" (UniqueName: \"kubernetes.io/projected/7bedc55a-2396-4249-ac54-a753a67dca72-kube-api-access-z5mws\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194321 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-secrets\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194349 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-generated\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.194383 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rngf\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-kube-api-access-8rngf\") pod \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.195803 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.195928 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.196199 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-combined-ca-bundle\") pod \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.196234 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-public-tls-certs\") pod \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.196274 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.196313 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-run-httpd\") pod \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.196353 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-etc-swift\") pod \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\" (UID: \"51e99221-24f0-4cbe-bb3c-4e1e70acd4ba\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.196414 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-default\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.196479 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-kolla-config\") pod \"7bedc55a-2396-4249-ac54-a753a67dca72\" (UID: \"7bedc55a-2396-4249-ac54-a753a67dca72\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.196480 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" (UID: "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.200439 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.201128 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" (UID: "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.201885 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.202122 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.202264 4935 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.202278 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.202290 4935 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bedc55a-2396-4249-ac54-a753a67dca72-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.202306 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7fqw\" (UniqueName: \"kubernetes.io/projected/d6ccd5ed-23d9-4d0b-99fb-68655c0199b5-kube-api-access-l7fqw\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.202319 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7bedc55a-2396-4249-ac54-a753a67dca72-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.202334 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.224410 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-secrets" (OuterVolumeSpecName: "secrets") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.246845 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-kube-api-access-8rngf" (OuterVolumeSpecName: "kube-api-access-8rngf") pod "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" (UID: "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba"). InnerVolumeSpecName "kube-api-access-8rngf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.280672 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bedc55a-2396-4249-ac54-a753a67dca72-kube-api-access-z5mws" (OuterVolumeSpecName: "kube-api-access-z5mws") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "kube-api-access-z5mws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.281365 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.296040 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" (UID: "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322579 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-certs\") pod \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322639 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9jml\" (UniqueName: \"kubernetes.io/projected/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-api-access-d9jml\") pod \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322669 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data\") pod \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322722 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9acce3fe-5a42-49e7-b3e7-a72f01af5736-etc-machine-id\") pod \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322775 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-combined-ca-bundle\") pod \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322870 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data-custom\") pod \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322904 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rjgb\" (UniqueName: \"kubernetes.io/projected/9acce3fe-5a42-49e7-b3e7-a72f01af5736-kube-api-access-2rjgb\") pod \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322922 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-combined-ca-bundle\") pod \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.322971 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-config\") pod \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\" (UID: \"c710b93b-c2fe-4a03-ba8c-f280a9d67da6\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.323011 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-scripts\") pod \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\" (UID: \"9acce3fe-5a42-49e7-b3e7-a72f01af5736\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.324192 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5mws\" (UniqueName: \"kubernetes.io/projected/7bedc55a-2396-4249-ac54-a753a67dca72-kube-api-access-z5mws\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.324211 4935 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.324219 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rngf\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-kube-api-access-8rngf\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.324227 4935 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.334547 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.338993 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-scripts" (OuterVolumeSpecName: "scripts") pod "9acce3fe-5a42-49e7-b3e7-a72f01af5736" (UID: "9acce3fe-5a42-49e7-b3e7-a72f01af5736"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.351400 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9acce3fe-5a42-49e7-b3e7-a72f01af5736" (UID: "9acce3fe-5a42-49e7-b3e7-a72f01af5736"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.351632 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9acce3fe-5a42-49e7-b3e7-a72f01af5736-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9acce3fe-5a42-49e7-b3e7-a72f01af5736" (UID: "9acce3fe-5a42-49e7-b3e7-a72f01af5736"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.353228 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9acce3fe-5a42-49e7-b3e7-a72f01af5736-kube-api-access-2rjgb" (OuterVolumeSpecName: "kube-api-access-2rjgb") pod "9acce3fe-5a42-49e7-b3e7-a72f01af5736" (UID: "9acce3fe-5a42-49e7-b3e7-a72f01af5736"). InnerVolumeSpecName "kube-api-access-2rjgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.353381 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-api-access-d9jml" (OuterVolumeSpecName: "kube-api-access-d9jml") pod "c710b93b-c2fe-4a03-ba8c-f280a9d67da6" (UID: "c710b93b-c2fe-4a03-ba8c-f280a9d67da6"). InnerVolumeSpecName "kube-api-access-d9jml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.353372 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.424899 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-combined-ca-bundle\") pod \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.424955 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-internal-tls-certs\") pod \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.424978 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts\") pod \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425055 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-logs\") pod \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425186 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gp87c\" (UniqueName: \"kubernetes.io/projected/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-kube-api-access-gp87c\") pod \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425257 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data\") pod \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425316 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-public-tls-certs\") pod \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\" (UID: \"88a7ceab-efbb-4ba4-a5f2-c6702abbc314\") " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425621 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425655 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rjgb\" (UniqueName: \"kubernetes.io/projected/9acce3fe-5a42-49e7-b3e7-a72f01af5736-kube-api-access-2rjgb\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425678 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425690 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425700 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9jml\" (UniqueName: \"kubernetes.io/projected/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-api-access-d9jml\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425709 4935 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9acce3fe-5a42-49e7-b3e7-a72f01af5736-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.425717 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.426441 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-logs" (OuterVolumeSpecName: "logs") pod "88a7ceab-efbb-4ba4-a5f2-c6702abbc314" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.471791 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts" (OuterVolumeSpecName: "scripts") pod "88a7ceab-efbb-4ba4-a5f2-c6702abbc314" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.471822 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-kube-api-access-gp87c" (OuterVolumeSpecName: "kube-api-access-gp87c") pod "88a7ceab-efbb-4ba4-a5f2-c6702abbc314" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314"). InnerVolumeSpecName "kube-api-access-gp87c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.527634 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.527660 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gp87c\" (UniqueName: \"kubernetes.io/projected/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-kube-api-access-gp87c\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.527670 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.554497 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.578853 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" (UID: "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.599391 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "c710b93b-c2fe-4a03-ba8c-f280a9d67da6" (UID: "c710b93b-c2fe-4a03-ba8c-f280a9d67da6"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.603393 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "7bedc55a-2396-4249-ac54-a753a67dca72" (UID: "7bedc55a-2396-4249-ac54-a753a67dca72"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.607839 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-config-data" (OuterVolumeSpecName: "config-data") pod "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" (UID: "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.622382 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" (UID: "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.629987 4935 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.630031 4935 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bedc55a-2396-4249-ac54-a753a67dca72-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.630042 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.630051 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.630059 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.630067 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.664288 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "c710b93b-c2fe-4a03-ba8c-f280a9d67da6" (UID: "c710b93b-c2fe-4a03-ba8c-f280a9d67da6"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.667101 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancee563-account-delete-dqlrz" event={"ID":"6bd6e7de-59a5-479d-b2b8-7f060883f6c8","Type":"ContainerStarted","Data":"7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.667527 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glancee563-account-delete-dqlrz" podUID="6bd6e7de-59a5-479d-b2b8-7f060883f6c8" containerName="mariadb-account-delete" containerID="cri-o://7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7" gracePeriod=30 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.674260 4935 generic.go:334] "Generic (PLEG): container finished" podID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerID="37c00f2d3e9897fa787e5b2b4870c3b4e33ad0c59b0775763a7aeb443b6567b7" exitCode=0 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.674730 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerDied","Data":"37c00f2d3e9897fa787e5b2b4870c3b4e33ad0c59b0775763a7aeb443b6567b7"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.679009 4935 generic.go:334] "Generic (PLEG): container finished" podID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerID="f98ff5241a039270a3b05822fea3c87f18859f74162897e6acb797aeca39838d" exitCode=0 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.679159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74e20071-1ff2-4fae-ae8c-beb9dbce35ce","Type":"ContainerDied","Data":"f98ff5241a039270a3b05822fea3c87f18859f74162897e6acb797aeca39838d"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.684022 4935 generic.go:334] "Generic (PLEG): container finished" podID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerID="1ca4b6cdb5511dcb0a9782ff53d490a4721c69f45e3fc5128541c3b64c51ddd3" exitCode=0 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.684098 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c","Type":"ContainerDied","Data":"1ca4b6cdb5511dcb0a9782ff53d490a4721c69f45e3fc5128541c3b64c51ddd3"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.685971 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell09d1d-account-delete-24dn6" event={"ID":"7143940e-eedf-48a0-b0dc-4ec91df08ad0","Type":"ContainerStarted","Data":"027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.686080 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell09d1d-account-delete-24dn6" podUID="7143940e-eedf-48a0-b0dc-4ec91df08ad0" containerName="mariadb-account-delete" containerID="cri-o://027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9" gracePeriod=30 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.687980 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glancee563-account-delete-dqlrz" podStartSLOduration=6.687875049 podStartE2EDuration="6.687875049s" podCreationTimestamp="2025-10-05 07:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:15:54.681983694 +0000 UTC m=+1388.564610154" watchObservedRunningTime="2025-10-05 07:15:54.687875049 +0000 UTC m=+1388.570501509" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.693182 4935 generic.go:334] "Generic (PLEG): container finished" podID="b82213d6-93fa-48fc-808b-36f902cce15c" containerID="fb6f5493ca2adedadd805d4c56c29f8cf8470abc87c240d1db6707f07565dd60" exitCode=0 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.693310 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b82213d6-93fa-48fc-808b-36f902cce15c","Type":"ContainerDied","Data":"fb6f5493ca2adedadd805d4c56c29f8cf8470abc87c240d1db6707f07565dd60"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.699681 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88a7ceab-efbb-4ba4-a5f2-c6702abbc314" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.704380 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanb178-account-delete-rnqtn" event={"ID":"b732474b-e717-4d8d-91b7-e00409800f85","Type":"ContainerStarted","Data":"cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.704472 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbicanb178-account-delete-rnqtn" podUID="b732474b-e717-4d8d-91b7-e00409800f85" containerName="mariadb-account-delete" containerID="cri-o://cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd" gracePeriod=30 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.707649 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c710b93b-c2fe-4a03-ba8c-f280a9d67da6" (UID: "c710b93b-c2fe-4a03-ba8c-f280a9d67da6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.709543 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data" (OuterVolumeSpecName: "config-data") pod "88a7ceab-efbb-4ba4-a5f2-c6702abbc314" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.709576 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" (UID: "51e99221-24f0-4cbe-bb3c-4e1e70acd4ba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.721633 4935 generic.go:334] "Generic (PLEG): container finished" podID="6c512e7c-9418-43b4-8235-4d62fa487556" containerID="fc1c1a196142dfa9f9def44b621bd771b1f8fbde0b745b2687d81b122f138212" exitCode=0 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.721694 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c512e7c-9418-43b4-8235-4d62fa487556","Type":"ContainerDied","Data":"fc1c1a196142dfa9f9def44b621bd771b1f8fbde0b745b2687d81b122f138212"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.721721 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c512e7c-9418-43b4-8235-4d62fa487556","Type":"ContainerDied","Data":"a1ba0acdc2f2c344701c2fc076f77ad8459cad00319953c4929ff569e48699b4"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.721732 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1ba0acdc2f2c344701c2fc076f77ad8459cad00319953c4929ff569e48699b4" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.723017 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "88a7ceab-efbb-4ba4-a5f2-c6702abbc314" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.724050 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbdfd7554-jrf6q" event={"ID":"88a7ceab-efbb-4ba4-a5f2-c6702abbc314","Type":"ContainerDied","Data":"cf7c7ec1a50d259d4f557b17718ae497667bf183d921c9fe0b161a7bc0ba5a52"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.724101 4935 scope.go:117] "RemoveContainer" containerID="829086067b4c67ac2011dcae6d0f8b4b1d850acfb49e4cf8b466a30f81bb3be2" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.724201 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6dbdfd7554-jrf6q" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.727987 4935 generic.go:334] "Generic (PLEG): container finished" podID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerID="119888a24ef912962cb413c88d9903d21aad373b3de77a30cfb8ecc482ff8f75" exitCode=0 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.728035 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74989f54d8-bw287" event={"ID":"3c422d4a-a835-443a-9c6d-3161bc80c1aa","Type":"ContainerDied","Data":"119888a24ef912962cb413c88d9903d21aad373b3de77a30cfb8ecc482ff8f75"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.728054 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74989f54d8-bw287" event={"ID":"3c422d4a-a835-443a-9c6d-3161bc80c1aa","Type":"ContainerDied","Data":"c584480b35bceb761270e71b6c5a36b7d7a92b0e54a1575e6dbcf0164b50ed3b"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.728065 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c584480b35bceb761270e71b6c5a36b7d7a92b0e54a1575e6dbcf0164b50ed3b" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.730157 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder9bb0-account-delete-8m44f" event={"ID":"420a72a6-506a-42e6-b77a-ee5ca8527768","Type":"ContainerStarted","Data":"f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.730275 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder9bb0-account-delete-8m44f" podUID="420a72a6-506a-42e6-b77a-ee5ca8527768" containerName="mariadb-account-delete" containerID="cri-o://f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87" gracePeriod=30 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.734296 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.734315 4935 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.734325 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.734333 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.734341 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c710b93b-c2fe-4a03-ba8c-f280a9d67da6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.734349 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.735699 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell09d1d-account-delete-24dn6" podStartSLOduration=6.735677774 podStartE2EDuration="6.735677774s" podCreationTimestamp="2025-10-05 07:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:15:54.70035918 +0000 UTC m=+1388.582985640" watchObservedRunningTime="2025-10-05 07:15:54.735677774 +0000 UTC m=+1388.618304244" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.736945 4935 generic.go:334] "Generic (PLEG): container finished" podID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerID="573548ed93406199c78ab252ce72ca55ded60b3bb8509475fa7af985c3e26a02" exitCode=0 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.736987 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f1eb003-e111-4a86-b6c7-a899aa0426a0","Type":"ContainerDied","Data":"573548ed93406199c78ab252ce72ca55ded60b3bb8509475fa7af985c3e26a02"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.742491 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.743241 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.743575 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi70bb-account-delete-5wwbr" event={"ID":"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a","Type":"ContainerStarted","Data":"b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3"} Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.743783 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapi70bb-account-delete-5wwbr" podUID="819c9cf6-f5e5-46af-9fee-c1c4cc88c29a" containerName="mariadb-account-delete" containerID="cri-o://b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3" gracePeriod=30 Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.744177 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9acce3fe-5a42-49e7-b3e7-a72f01af5736" (UID: "9acce3fe-5a42-49e7-b3e7-a72f01af5736"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.744218 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.744233 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.744371 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6f6d895595-jv76b" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.746525 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb3bb-account-delete-v2qzp" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.749377 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbicanb178-account-delete-rnqtn" podStartSLOduration=6.749357176 podStartE2EDuration="6.749357176s" podCreationTimestamp="2025-10-05 07:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:15:54.720277857 +0000 UTC m=+1388.602904317" watchObservedRunningTime="2025-10-05 07:15:54.749357176 +0000 UTC m=+1388.631983636" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.762925 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder9bb0-account-delete-8m44f" podStartSLOduration=6.762882694 podStartE2EDuration="6.762882694s" podCreationTimestamp="2025-10-05 07:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:15:54.741133609 +0000 UTC m=+1388.623760069" watchObservedRunningTime="2025-10-05 07:15:54.762882694 +0000 UTC m=+1388.645509154" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.769855 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapi70bb-account-delete-5wwbr" podStartSLOduration=5.769844298 podStartE2EDuration="5.769844298s" podCreationTimestamp="2025-10-05 07:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:15:54.760379248 +0000 UTC m=+1388.643005708" watchObservedRunningTime="2025-10-05 07:15:54.769844298 +0000 UTC m=+1388.652470758" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.787723 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2967548c-91b1-43de-9c7d-53b462b09478" path="/var/lib/kubelet/pods/2967548c-91b1-43de-9c7d-53b462b09478/volumes" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.788301 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8" path="/var/lib/kubelet/pods/40c9b9aa-b26e-4b96-87a1-c7c82a1bffd8/volumes" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.788821 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" path="/var/lib/kubelet/pods/7c7acc7a-797f-46cd-b298-1f28fd2951d6/volumes" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.789836 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data" (OuterVolumeSpecName: "config-data") pod "9acce3fe-5a42-49e7-b3e7-a72f01af5736" (UID: "9acce3fe-5a42-49e7-b3e7-a72f01af5736"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.790185 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae9d561d-a53d-4b85-9232-776b137fb12c" path="/var/lib/kubelet/pods/ae9d561d-a53d-4b85-9232-776b137fb12c/volumes" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.790695 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b520d752-90d4-41a1-a6d0-243e7058c47c" path="/var/lib/kubelet/pods/b520d752-90d4-41a1-a6d0-243e7058c47c/volumes" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.791426 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98d30f9-d435-4c12-80c9-0f03a3853a0f" path="/var/lib/kubelet/pods/e98d30f9-d435-4c12-80c9-0f03a3853a0f/volumes" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.791807 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "88a7ceab-efbb-4ba4-a5f2-c6702abbc314" (UID: "88a7ceab-efbb-4ba4-a5f2-c6702abbc314"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.793746 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3fbc168-2981-4a9c-b4c4-0a6f1cec4556" path="/var/lib/kubelet/pods/f3fbc168-2981-4a9c-b4c4-0a6f1cec4556/volumes" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.836424 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.836452 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/88a7ceab-efbb-4ba4-a5f2-c6702abbc314-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.836464 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acce3fe-5a42-49e7-b3e7-a72f01af5736-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.915868 4935 scope.go:117] "RemoveContainer" containerID="27997ed35ed34c0a762315ae80ed4c65e006d4cd919243a41130b860e5a4e10a" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.963130 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.971355 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:15:54 crc kubenswrapper[4935]: I1005 07:15:54.993671 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.028221 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.040066 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-internal-tls-certs\") pod \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.040096 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data-custom\") pod \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.040413 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c422d4a-a835-443a-9c6d-3161bc80c1aa-logs\") pod \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.040479 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-combined-ca-bundle\") pod \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.040513 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data\") pod \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.040578 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-public-tls-certs\") pod \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.040639 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8dcg\" (UniqueName: \"kubernetes.io/projected/3c422d4a-a835-443a-9c6d-3161bc80c1aa-kube-api-access-d8dcg\") pod \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\" (UID: \"3c422d4a-a835-443a-9c6d-3161bc80c1aa\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.041119 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c422d4a-a835-443a-9c6d-3161bc80c1aa-logs" (OuterVolumeSpecName: "logs") pod "3c422d4a-a835-443a-9c6d-3161bc80c1aa" (UID: "3c422d4a-a835-443a-9c6d-3161bc80c1aa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.055864 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3c422d4a-a835-443a-9c6d-3161bc80c1aa" (UID: "3c422d4a-a835-443a-9c6d-3161bc80c1aa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.061454 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.066262 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c422d4a-a835-443a-9c6d-3161bc80c1aa-kube-api-access-d8dcg" (OuterVolumeSpecName: "kube-api-access-d8dcg") pod "3c422d4a-a835-443a-9c6d-3161bc80c1aa" (UID: "3c422d4a-a835-443a-9c6d-3161bc80c1aa"). InnerVolumeSpecName "kube-api-access-d8dcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.075697 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.111222 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb3bb-account-delete-v2qzp"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.127094 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c422d4a-a835-443a-9c6d-3161bc80c1aa" (UID: "3c422d4a-a835-443a-9c6d-3161bc80c1aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.128072 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronb3bb-account-delete-v2qzp"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148706 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-config-data\") pod \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148746 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-nova-metadata-tls-certs\") pod \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148787 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-logs\") pod \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148820 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-ceilometer-tls-certs\") pod \"aba39a0b-f275-4075-9419-a354a3c5fa5e\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148834 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-public-tls-certs\") pod \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148849 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-combined-ca-bundle\") pod \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148868 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-log-httpd\") pod \"aba39a0b-f275-4075-9419-a354a3c5fa5e\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148906 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-internal-tls-certs\") pod \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148933 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-combined-ca-bundle\") pod \"aba39a0b-f275-4075-9419-a354a3c5fa5e\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148952 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-scripts\") pod \"6c512e7c-9418-43b4-8235-4d62fa487556\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148972 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-public-tls-certs\") pod \"6c512e7c-9418-43b4-8235-4d62fa487556\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.148988 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-logs\") pod \"6c512e7c-9418-43b4-8235-4d62fa487556\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149010 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-config-data\") pod \"aba39a0b-f275-4075-9419-a354a3c5fa5e\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149023 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-config-data\") pod \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149052 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtgv9\" (UniqueName: \"kubernetes.io/projected/9f1eb003-e111-4a86-b6c7-a899aa0426a0-kube-api-access-vtgv9\") pod \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149076 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-config-data\") pod \"6c512e7c-9418-43b4-8235-4d62fa487556\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149100 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-scripts\") pod \"aba39a0b-f275-4075-9419-a354a3c5fa5e\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149114 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-run-httpd\") pod \"aba39a0b-f275-4075-9419-a354a3c5fa5e\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149132 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1eb003-e111-4a86-b6c7-a899aa0426a0-logs\") pod \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149177 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf4f5\" (UniqueName: \"kubernetes.io/projected/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-kube-api-access-bf4f5\") pod \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\" (UID: \"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149239 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-combined-ca-bundle\") pod \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\" (UID: \"9f1eb003-e111-4a86-b6c7-a899aa0426a0\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149261 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-combined-ca-bundle\") pod \"6c512e7c-9418-43b4-8235-4d62fa487556\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149286 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs7jr\" (UniqueName: \"kubernetes.io/projected/aba39a0b-f275-4075-9419-a354a3c5fa5e-kube-api-access-rs7jr\") pod \"aba39a0b-f275-4075-9419-a354a3c5fa5e\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149302 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-sg-core-conf-yaml\") pod \"aba39a0b-f275-4075-9419-a354a3c5fa5e\" (UID: \"aba39a0b-f275-4075-9419-a354a3c5fa5e\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149319 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-httpd-run\") pod \"6c512e7c-9418-43b4-8235-4d62fa487556\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149335 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ssrz\" (UniqueName: \"kubernetes.io/projected/6c512e7c-9418-43b4-8235-4d62fa487556-kube-api-access-8ssrz\") pod \"6c512e7c-9418-43b4-8235-4d62fa487556\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149351 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"6c512e7c-9418-43b4-8235-4d62fa487556\" (UID: \"6c512e7c-9418-43b4-8235-4d62fa487556\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149643 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c422d4a-a835-443a-9c6d-3161bc80c1aa-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149663 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149673 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8dcg\" (UniqueName: \"kubernetes.io/projected/3c422d4a-a835-443a-9c6d-3161bc80c1aa-kube-api-access-d8dcg\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.149681 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.150524 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3c422d4a-a835-443a-9c6d-3161bc80c1aa" (UID: "3c422d4a-a835-443a-9c6d-3161bc80c1aa"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.153446 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "6c512e7c-9418-43b4-8235-4d62fa487556" (UID: "6c512e7c-9418-43b4-8235-4d62fa487556"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.155516 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aba39a0b-f275-4075-9419-a354a3c5fa5e-kube-api-access-rs7jr" (OuterVolumeSpecName: "kube-api-access-rs7jr") pod "aba39a0b-f275-4075-9419-a354a3c5fa5e" (UID: "aba39a0b-f275-4075-9419-a354a3c5fa5e"). InnerVolumeSpecName "kube-api-access-rs7jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.159598 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6c512e7c-9418-43b4-8235-4d62fa487556" (UID: "6c512e7c-9418-43b4-8235-4d62fa487556"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.162291 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-kube-api-access-bf4f5" (OuterVolumeSpecName: "kube-api-access-bf4f5") pod "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" (UID: "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c"). InnerVolumeSpecName "kube-api-access-bf4f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.162435 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-scripts" (OuterVolumeSpecName: "scripts") pod "aba39a0b-f275-4075-9419-a354a3c5fa5e" (UID: "aba39a0b-f275-4075-9419-a354a3c5fa5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.162576 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aba39a0b-f275-4075-9419-a354a3c5fa5e" (UID: "aba39a0b-f275-4075-9419-a354a3c5fa5e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.162865 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c512e7c-9418-43b4-8235-4d62fa487556-kube-api-access-8ssrz" (OuterVolumeSpecName: "kube-api-access-8ssrz") pod "6c512e7c-9418-43b4-8235-4d62fa487556" (UID: "6c512e7c-9418-43b4-8235-4d62fa487556"). InnerVolumeSpecName "kube-api-access-8ssrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.163023 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f1eb003-e111-4a86-b6c7-a899aa0426a0-logs" (OuterVolumeSpecName: "logs") pod "9f1eb003-e111-4a86-b6c7-a899aa0426a0" (UID: "9f1eb003-e111-4a86-b6c7-a899aa0426a0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.163476 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.164091 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-logs" (OuterVolumeSpecName: "logs") pod "6c512e7c-9418-43b4-8235-4d62fa487556" (UID: "6c512e7c-9418-43b4-8235-4d62fa487556"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.164180 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aba39a0b-f275-4075-9419-a354a3c5fa5e" (UID: "aba39a0b-f275-4075-9419-a354a3c5fa5e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.164977 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-logs" (OuterVolumeSpecName: "logs") pod "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" (UID: "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.164987 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.169172 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.204004 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-scripts" (OuterVolumeSpecName: "scripts") pod "6c512e7c-9418-43b4-8235-4d62fa487556" (UID: "6c512e7c-9418-43b4-8235-4d62fa487556"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.204263 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f1eb003-e111-4a86-b6c7-a899aa0426a0-kube-api-access-vtgv9" (OuterVolumeSpecName: "kube-api-access-vtgv9") pod "9f1eb003-e111-4a86-b6c7-a899aa0426a0" (UID: "9f1eb003-e111-4a86-b6c7-a899aa0426a0"). InnerVolumeSpecName "kube-api-access-vtgv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.220298 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6f6d895595-jv76b"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.237806 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-6f6d895595-jv76b"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.248280 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3c422d4a-a835-443a-9c6d-3161bc80c1aa" (UID: "3c422d4a-a835-443a-9c6d-3161bc80c1aa"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.248338 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data" (OuterVolumeSpecName: "config-data") pod "3c422d4a-a835-443a-9c6d-3161bc80c1aa" (UID: "3c422d4a-a835-443a-9c6d-3161bc80c1aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.256616 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-config-data\") pod \"b82213d6-93fa-48fc-808b-36f902cce15c\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.256715 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cw9j\" (UniqueName: \"kubernetes.io/projected/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-kube-api-access-7cw9j\") pod \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.256760 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-combined-ca-bundle\") pod \"b82213d6-93fa-48fc-808b-36f902cce15c\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.257098 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-config-data\") pod \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.257221 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-memcached-tls-certs\") pod \"b82213d6-93fa-48fc-808b-36f902cce15c\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.257280 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wn64\" (UniqueName: \"kubernetes.io/projected/b82213d6-93fa-48fc-808b-36f902cce15c-kube-api-access-6wn64\") pod \"b82213d6-93fa-48fc-808b-36f902cce15c\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.257323 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-kolla-config\") pod \"b82213d6-93fa-48fc-808b-36f902cce15c\" (UID: \"b82213d6-93fa-48fc-808b-36f902cce15c\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.257373 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-combined-ca-bundle\") pod \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\" (UID: \"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258114 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258131 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258146 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258155 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258165 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258175 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtgv9\" (UniqueName: \"kubernetes.io/projected/9f1eb003-e111-4a86-b6c7-a899aa0426a0-kube-api-access-vtgv9\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258187 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aba39a0b-f275-4075-9419-a354a3c5fa5e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258196 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258206 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1eb003-e111-4a86-b6c7-a899aa0426a0-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258214 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258226 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf4f5\" (UniqueName: \"kubernetes.io/projected/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-kube-api-access-bf4f5\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258234 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c422d4a-a835-443a-9c6d-3161bc80c1aa-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258243 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs7jr\" (UniqueName: \"kubernetes.io/projected/aba39a0b-f275-4075-9419-a354a3c5fa5e-kube-api-access-rs7jr\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258252 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c512e7c-9418-43b4-8235-4d62fa487556-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258263 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ssrz\" (UniqueName: \"kubernetes.io/projected/6c512e7c-9418-43b4-8235-4d62fa487556-kube-api-access-8ssrz\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.258286 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.260343 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-config-data" (OuterVolumeSpecName: "config-data") pod "b82213d6-93fa-48fc-808b-36f902cce15c" (UID: "b82213d6-93fa-48fc-808b-36f902cce15c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.261228 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.261285 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.264034 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "b82213d6-93fa-48fc-808b-36f902cce15c" (UID: "b82213d6-93fa-48fc-808b-36f902cce15c"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.275755 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-config-data" (OuterVolumeSpecName: "config-data") pod "9f1eb003-e111-4a86-b6c7-a899aa0426a0" (UID: "9f1eb003-e111-4a86-b6c7-a899aa0426a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.282272 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-kube-api-access-7cw9j" (OuterVolumeSpecName: "kube-api-access-7cw9j") pod "d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" (UID: "d6d217a2-8ab3-46a8-a70b-58cb5d5cad32"). InnerVolumeSpecName "kube-api-access-7cw9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.293843 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b82213d6-93fa-48fc-808b-36f902cce15c-kube-api-access-6wn64" (OuterVolumeSpecName: "kube-api-access-6wn64") pod "b82213d6-93fa-48fc-808b-36f902cce15c" (UID: "b82213d6-93fa-48fc-808b-36f902cce15c"). InnerVolumeSpecName "kube-api-access-6wn64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.342372 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-config-data" (OuterVolumeSpecName: "config-data") pod "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" (UID: "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.343055 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.359653 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.359695 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wn64\" (UniqueName: \"kubernetes.io/projected/b82213d6-93fa-48fc-808b-36f902cce15c-kube-api-access-6wn64\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.359721 4935 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.359732 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.359740 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82213d6-93fa-48fc-808b-36f902cce15c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.359748 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cw9j\" (UniqueName: \"kubernetes.io/projected/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-kube-api-access-7cw9j\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.359757 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.367756 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6dbdfd7554-jrf6q"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.369384 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aba39a0b-f275-4075-9419-a354a3c5fa5e" (UID: "aba39a0b-f275-4075-9419-a354a3c5fa5e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.369942 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" (UID: "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.371081 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6c512e7c-9418-43b4-8235-4d62fa487556" (UID: "6c512e7c-9418-43b4-8235-4d62fa487556"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.374429 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6dbdfd7554-jrf6q"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.380536 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-config-data" (OuterVolumeSpecName: "config-data") pod "6c512e7c-9418-43b4-8235-4d62fa487556" (UID: "6c512e7c-9418-43b4-8235-4d62fa487556"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.385346 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.392603 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.399438 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" (UID: "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.407700 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.409334 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f1eb003-e111-4a86-b6c7-a899aa0426a0" (UID: "9f1eb003-e111-4a86-b6c7-a899aa0426a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.464112 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.464382 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.464391 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.464403 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.464411 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.464419 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.464427 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.487183 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" (UID: "d6d217a2-8ab3-46a8-a70b-58cb5d5cad32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.492024 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" (UID: "e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.493664 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b82213d6-93fa-48fc-808b-36f902cce15c" (UID: "b82213d6-93fa-48fc-808b-36f902cce15c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.496562 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9f1eb003-e111-4a86-b6c7-a899aa0426a0" (UID: "9f1eb003-e111-4a86-b6c7-a899aa0426a0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.518011 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "b82213d6-93fa-48fc-808b-36f902cce15c" (UID: "b82213d6-93fa-48fc-808b-36f902cce15c"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.523868 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c512e7c-9418-43b4-8235-4d62fa487556" (UID: "6c512e7c-9418-43b4-8235-4d62fa487556"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.524432 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aba39a0b-f275-4075-9419-a354a3c5fa5e" (UID: "aba39a0b-f275-4075-9419-a354a3c5fa5e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.526090 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-config-data" (OuterVolumeSpecName: "config-data") pod "aba39a0b-f275-4075-9419-a354a3c5fa5e" (UID: "aba39a0b-f275-4075-9419-a354a3c5fa5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.537873 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-config-data" (OuterVolumeSpecName: "config-data") pod "d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" (UID: "d6d217a2-8ab3-46a8-a70b-58cb5d5cad32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.551082 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aba39a0b-f275-4075-9419-a354a3c5fa5e" (UID: "aba39a0b-f275-4075-9419-a354a3c5fa5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566291 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c512e7c-9418-43b4-8235-4d62fa487556-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566322 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566332 4935 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1eb003-e111-4a86-b6c7-a899aa0426a0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566343 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566352 4935 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566360 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566370 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566378 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566385 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba39a0b-f275-4075-9419-a354a3c5fa5e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.566393 4935 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b82213d6-93fa-48fc-808b-36f902cce15c-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: E1005 07:15:55.660369 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6 is running failed: container process not found" containerID="1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:15:55 crc kubenswrapper[4935]: E1005 07:15:55.660654 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6 is running failed: container process not found" containerID="1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:15:55 crc kubenswrapper[4935]: E1005 07:15:55.660982 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6 is running failed: container process not found" containerID="1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:15:55 crc kubenswrapper[4935]: E1005 07:15:55.661007 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="ovn-northd" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.700358 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.703078 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.776085 4935 generic.go:334] "Generic (PLEG): container finished" podID="d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" containerID="2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f" exitCode=0 Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.776408 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32","Type":"ContainerDied","Data":"2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.776438 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6d217a2-8ab3-46a8-a70b-58cb5d5cad32","Type":"ContainerDied","Data":"76c1107eef30a0a722711178233696fe10f16a1b12c802573671e8326b03cc7e"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.776455 4935 scope.go:117] "RemoveContainer" containerID="2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.776595 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.782552 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aba39a0b-f275-4075-9419-a354a3c5fa5e","Type":"ContainerDied","Data":"cfed2e35aa6cff13d197095c51fbf6bf35c8a059124e7708c37f3b502f0eed87"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.782656 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.794165 4935 generic.go:334] "Generic (PLEG): container finished" podID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" containerID="09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d" exitCode=0 Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.794277 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"af041b3a-6ed3-4c8d-836b-a9b69f95360b","Type":"ContainerDied","Data":"09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.794308 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"af041b3a-6ed3-4c8d-836b-a9b69f95360b","Type":"ContainerDied","Data":"8496b167da6a6e05fbdffbb870edca983c3b713a036855854905138de03c1df0"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.794389 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.811075 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.811140 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74e20071-1ff2-4fae-ae8c-beb9dbce35ce","Type":"ContainerDied","Data":"029b36e747b3b514a2e5123e18003f92080faaf01ed4d2dc433120d5adcbe579"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.817521 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.829126 4935 scope.go:117] "RemoveContainer" containerID="2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.829423 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7c412f4e-54aa-4d6c-9606-753f97474c3b/ovn-northd/0.log" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.829468 4935 generic.go:334] "Generic (PLEG): container finished" podID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerID="1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" exitCode=139 Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.829523 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7c412f4e-54aa-4d6c-9606-753f97474c3b","Type":"ContainerDied","Data":"1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6"} Oct 05 07:15:55 crc kubenswrapper[4935]: E1005 07:15:55.830710 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f\": container with ID starting with 2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f not found: ID does not exist" containerID="2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.830771 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f"} err="failed to get container status \"2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f\": rpc error: code = NotFound desc = could not find container \"2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f\": container with ID starting with 2304aaf640adb363ca16e7a8a1d17840e37240d2dc11030600eafe9f73a8501f not found: ID does not exist" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.830799 4935 scope.go:117] "RemoveContainer" containerID="fc4511ccd783fd2b8c6cb4dce5c4f3cc88512854346a9a9f11bfe9081fb534e4" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.834938 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b82213d6-93fa-48fc-808b-36f902cce15c","Type":"ContainerDied","Data":"325c92c64a0778df03cd0010f03519dc08238f80cfa11b3b7fbcc352581236cb"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.835079 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.835146 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.841741 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.844290 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f1eb003-e111-4a86-b6c7-a899aa0426a0","Type":"ContainerDied","Data":"a26d2a64080f031f3f81dc5b5a40e927c359a81f36fa092010bad367d21a0dc1"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.844382 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.846862 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.856433 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.856602 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c","Type":"ContainerDied","Data":"ce69b9b67f63c1eabab425a4a92471582bea87b0eac36b58ede89a2347a51ed3"} Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.856746 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74989f54d8-bw287" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.858223 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.867606 4935 scope.go:117] "RemoveContainer" containerID="48ae75168e8a69d6f8a97da4dacb4141e0980a9ae4a1946601d92425b5f4f413" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.870599 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-config-data\") pod \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.870680 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-secrets\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.870768 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kolla-config\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.870796 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-internal-tls-certs\") pod \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.870841 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx9rd\" (UniqueName: \"kubernetes.io/projected/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-kube-api-access-cx9rd\") pod \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.870868 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-combined-ca-bundle\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871061 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-generated\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871121 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-httpd-run\") pod \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871151 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871176 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmm7l\" (UniqueName: \"kubernetes.io/projected/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kube-api-access-lmm7l\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871197 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-operator-scripts\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871225 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-default\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871251 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-logs\") pod \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871281 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871321 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-scripts\") pod \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871393 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-galera-tls-certs\") pod \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\" (UID: \"af041b3a-6ed3-4c8d-836b-a9b69f95360b\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871436 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-combined-ca-bundle\") pod \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\" (UID: \"74e20071-1ff2-4fae-ae8c-beb9dbce35ce\") " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.871581 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.872289 4935 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.872371 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.872555 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-logs" (OuterVolumeSpecName: "logs") pod "74e20071-1ff2-4fae-ae8c-beb9dbce35ce" (UID: "74e20071-1ff2-4fae-ae8c-beb9dbce35ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.873187 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.874572 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.876999 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "74e20071-1ff2-4fae-ae8c-beb9dbce35ce" (UID: "74e20071-1ff2-4fae-ae8c-beb9dbce35ce"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.877557 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "74e20071-1ff2-4fae-ae8c-beb9dbce35ce" (UID: "74e20071-1ff2-4fae-ae8c-beb9dbce35ce"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.895667 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-kube-api-access-cx9rd" (OuterVolumeSpecName: "kube-api-access-cx9rd") pod "74e20071-1ff2-4fae-ae8c-beb9dbce35ce" (UID: "74e20071-1ff2-4fae-ae8c-beb9dbce35ce"). InnerVolumeSpecName "kube-api-access-cx9rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.902925 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.907127 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kube-api-access-lmm7l" (OuterVolumeSpecName: "kube-api-access-lmm7l") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "kube-api-access-lmm7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.907283 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-secrets" (OuterVolumeSpecName: "secrets") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.910059 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.910996 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-scripts" (OuterVolumeSpecName: "scripts") pod "74e20071-1ff2-4fae-ae8c-beb9dbce35ce" (UID: "74e20071-1ff2-4fae-ae8c-beb9dbce35ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.943188 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.953412 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.962366 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.965748 4935 scope.go:117] "RemoveContainer" containerID="37c00f2d3e9897fa787e5b2b4870c3b4e33ad0c59b0775763a7aeb443b6567b7" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973814 4935 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973842 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmm7l\" (UniqueName: \"kubernetes.io/projected/af041b3a-6ed3-4c8d-836b-a9b69f95360b-kube-api-access-lmm7l\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973854 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973863 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973883 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973908 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973917 4935 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973926 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx9rd\" (UniqueName: \"kubernetes.io/projected/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-kube-api-access-cx9rd\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973934 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973941 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/af041b3a-6ed3-4c8d-836b-a9b69f95360b-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973949 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.973962 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.977275 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.981494 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-config-data" (OuterVolumeSpecName: "config-data") pod "74e20071-1ff2-4fae-ae8c-beb9dbce35ce" (UID: "74e20071-1ff2-4fae-ae8c-beb9dbce35ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.992017 4935 scope.go:117] "RemoveContainer" containerID="1a7bd0c309de5a9b869370ed377d8a46c8a2f0c566fa6724d9ed2c87fbfb0e4a" Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.992144 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74989f54d8-bw287"] Oct 05 07:15:55 crc kubenswrapper[4935]: I1005 07:15:55.998268 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.000801 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-74989f54d8-bw287"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.001010 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74e20071-1ff2-4fae-ae8c-beb9dbce35ce" (UID: "74e20071-1ff2-4fae-ae8c-beb9dbce35ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.012246 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.016288 4935 scope.go:117] "RemoveContainer" containerID="09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.019806 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.020025 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "74e20071-1ff2-4fae-ae8c-beb9dbce35ce" (UID: "74e20071-1ff2-4fae-ae8c-beb9dbce35ce"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.022358 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "af041b3a-6ed3-4c8d-836b-a9b69f95360b" (UID: "af041b3a-6ed3-4c8d-836b-a9b69f95360b"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.025945 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.037942 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.039116 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.049824 4935 scope.go:117] "RemoveContainer" containerID="7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.068799 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7c412f4e-54aa-4d6c-9606-753f97474c3b/ovn-northd/0.log" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.068866 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.075992 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.076022 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.076032 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74e20071-1ff2-4fae-ae8c-beb9dbce35ce-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.076040 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.076051 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.076059 4935 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/af041b3a-6ed3-4c8d-836b-a9b69f95360b-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.077182 4935 scope.go:117] "RemoveContainer" containerID="09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d" Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.077499 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d\": container with ID starting with 09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d not found: ID does not exist" containerID="09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.077527 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d"} err="failed to get container status \"09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d\": rpc error: code = NotFound desc = could not find container \"09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d\": container with ID starting with 09c864a6294a5240409705dfb44966883dc1f1b4b3972cd27710638db608134d not found: ID does not exist" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.077548 4935 scope.go:117] "RemoveContainer" containerID="7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f" Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.077787 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f\": container with ID starting with 7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f not found: ID does not exist" containerID="7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.077809 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f"} err="failed to get container status \"7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f\": rpc error: code = NotFound desc = could not find container \"7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f\": container with ID starting with 7928c9458c9a1e9a78bc2f80b21545f568d8f4b117400927325e15f77f9b551f not found: ID does not exist" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.077821 4935 scope.go:117] "RemoveContainer" containerID="f98ff5241a039270a3b05822fea3c87f18859f74162897e6acb797aeca39838d" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.097149 4935 scope.go:117] "RemoveContainer" containerID="af59bf08d75bada41121a63c5849541082f4196ebb916ac86beff7f230324a9b" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.134995 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.139015 4935 scope.go:117] "RemoveContainer" containerID="fb6f5493ca2adedadd805d4c56c29f8cf8470abc87c240d1db6707f07565dd60" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.140447 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.165515 4935 scope.go:117] "RemoveContainer" containerID="573548ed93406199c78ab252ce72ca55ded60b3bb8509475fa7af985c3e26a02" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.167114 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.178978 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-config\") pod \"7c412f4e-54aa-4d6c-9606-753f97474c3b\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.179049 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp5d6\" (UniqueName: \"kubernetes.io/projected/7c412f4e-54aa-4d6c-9606-753f97474c3b-kube-api-access-sp5d6\") pod \"7c412f4e-54aa-4d6c-9606-753f97474c3b\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.179185 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-scripts\") pod \"7c412f4e-54aa-4d6c-9606-753f97474c3b\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.179230 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-combined-ca-bundle\") pod \"7c412f4e-54aa-4d6c-9606-753f97474c3b\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.179265 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-metrics-certs-tls-certs\") pod \"7c412f4e-54aa-4d6c-9606-753f97474c3b\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.179310 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-rundir\") pod \"7c412f4e-54aa-4d6c-9606-753f97474c3b\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.179354 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-northd-tls-certs\") pod \"7c412f4e-54aa-4d6c-9606-753f97474c3b\" (UID: \"7c412f4e-54aa-4d6c-9606-753f97474c3b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.180035 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.180094 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data podName:8dcf22e2-2acd-4b49-a2bc-813a50f89409 nodeName:}" failed. No retries permitted until 2025-10-05 07:16:04.180078199 +0000 UTC m=+1398.062704659 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data") pod "rabbitmq-server-0" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409") : configmap "rabbitmq-config-data" not found Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.180459 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "7c412f4e-54aa-4d6c-9606-753f97474c3b" (UID: "7c412f4e-54aa-4d6c-9606-753f97474c3b"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.180833 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-config" (OuterVolumeSpecName: "config") pod "7c412f4e-54aa-4d6c-9606-753f97474c3b" (UID: "7c412f4e-54aa-4d6c-9606-753f97474c3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.181220 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-scripts" (OuterVolumeSpecName: "scripts") pod "7c412f4e-54aa-4d6c-9606-753f97474c3b" (UID: "7c412f4e-54aa-4d6c-9606-753f97474c3b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.181506 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.183191 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c412f4e-54aa-4d6c-9606-753f97474c3b-kube-api-access-sp5d6" (OuterVolumeSpecName: "kube-api-access-sp5d6") pod "7c412f4e-54aa-4d6c-9606-753f97474c3b" (UID: "7c412f4e-54aa-4d6c-9606-753f97474c3b"). InnerVolumeSpecName "kube-api-access-sp5d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.192703 4935 scope.go:117] "RemoveContainer" containerID="122b7e3178e70d3a5f9d4ca95b47f2e8787d8b3c771046206b29a4dc71f41a66" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.235690 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c412f4e-54aa-4d6c-9606-753f97474c3b" (UID: "7c412f4e-54aa-4d6c-9606-753f97474c3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.248909 4935 scope.go:117] "RemoveContainer" containerID="1ca4b6cdb5511dcb0a9782ff53d490a4721c69f45e3fc5128541c3b64c51ddd3" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.270138 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7c412f4e-54aa-4d6c-9606-753f97474c3b" (UID: "7c412f4e-54aa-4d6c-9606-753f97474c3b"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.271077 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "7c412f4e-54aa-4d6c-9606-753f97474c3b" (UID: "7c412f4e-54aa-4d6c-9606-753f97474c3b"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.277104 4935 scope.go:117] "RemoveContainer" containerID="3e0b1900e020511c33b937275937ce79dc33f680831c304611354b614ac0d6a6" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.280883 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp5d6\" (UniqueName: \"kubernetes.io/projected/7c412f4e-54aa-4d6c-9606-753f97474c3b-kube-api-access-sp5d6\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.280957 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.280969 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.280977 4935 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.280986 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.280995 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c412f4e-54aa-4d6c-9606-753f97474c3b-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.281027 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c412f4e-54aa-4d6c-9606-753f97474c3b-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.366997 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24a61e5d809f2202e8aff2ffc919b4f809f3ea3e05132213aaafa01f0e1fc0d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.368844 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24a61e5d809f2202e8aff2ffc919b4f809f3ea3e05132213aaafa01f0e1fc0d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.370168 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="24a61e5d809f2202e8aff2ffc919b4f809f3ea3e05132213aaafa01f0e1fc0d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.370238 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" containerName="nova-cell0-conductor-conductor" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.463770 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.195:5353: i/o timeout" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.729963 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.789936 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" path="/var/lib/kubelet/pods/3c422d4a-a835-443a-9c6d-3161bc80c1aa/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.790791 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" path="/var/lib/kubelet/pods/51e99221-24f0-4cbe-bb3c-4e1e70acd4ba/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.791501 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" path="/var/lib/kubelet/pods/6c512e7c-9418-43b4-8235-4d62fa487556/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.793612 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" path="/var/lib/kubelet/pods/74e20071-1ff2-4fae-ae8c-beb9dbce35ce/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.794166 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77853c76-2bbe-42e5-a0ef-03d5c5fe2402" path="/var/lib/kubelet/pods/77853c76-2bbe-42e5-a0ef-03d5c5fe2402/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.796170 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bedc55a-2396-4249-ac54-a753a67dca72" path="/var/lib/kubelet/pods/7bedc55a-2396-4249-ac54-a753a67dca72/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.796799 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" path="/var/lib/kubelet/pods/88a7ceab-efbb-4ba4-a5f2-c6702abbc314/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.797421 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" path="/var/lib/kubelet/pods/9acce3fe-5a42-49e7-b3e7-a72f01af5736/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.798566 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" path="/var/lib/kubelet/pods/9f1eb003-e111-4a86-b6c7-a899aa0426a0/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.799147 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" path="/var/lib/kubelet/pods/aba39a0b-f275-4075-9419-a354a3c5fa5e/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.800057 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" path="/var/lib/kubelet/pods/af041b3a-6ed3-4c8d-836b-a9b69f95360b/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.800656 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b82213d6-93fa-48fc-808b-36f902cce15c" path="/var/lib/kubelet/pods/b82213d6-93fa-48fc-808b-36f902cce15c/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.801161 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c710b93b-c2fe-4a03-ba8c-f280a9d67da6" path="/var/lib/kubelet/pods/c710b93b-c2fe-4a03-ba8c-f280a9d67da6/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.802043 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6ccd5ed-23d9-4d0b-99fb-68655c0199b5" path="/var/lib/kubelet/pods/d6ccd5ed-23d9-4d0b-99fb-68655c0199b5/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.802500 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" path="/var/lib/kubelet/pods/d6d217a2-8ab3-46a8-a70b-58cb5d5cad32/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.803026 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" path="/var/lib/kubelet/pods/e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c/volumes" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.871408 4935 generic.go:334] "Generic (PLEG): container finished" podID="6214d810-93b4-4537-a655-642df573048b" containerID="010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413" exitCode=0 Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.871462 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79d9bc6587-tvhv2" event={"ID":"6214d810-93b4-4537-a655-642df573048b","Type":"ContainerDied","Data":"010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413"} Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.871487 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79d9bc6587-tvhv2" event={"ID":"6214d810-93b4-4537-a655-642df573048b","Type":"ContainerDied","Data":"2ddd8c31b1bc7ebcb2278e234b4f7aa4d87109bcbd539f6f21dc2ef468d8bc2e"} Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.871503 4935 scope.go:117] "RemoveContainer" containerID="010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.871497 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79d9bc6587-tvhv2" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.874464 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7c412f4e-54aa-4d6c-9606-753f97474c3b/ovn-northd/0.log" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.874541 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7c412f4e-54aa-4d6c-9606-753f97474c3b","Type":"ContainerDied","Data":"f4fa02e54a41f112ed3743bd55ccac348e9e982e4645e9e8085878490221d327"} Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.874638 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.897619 4935 scope.go:117] "RemoveContainer" containerID="010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413" Oct 05 07:15:56 crc kubenswrapper[4935]: E1005 07:15:56.897956 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413\": container with ID starting with 010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413 not found: ID does not exist" containerID="010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.897987 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413"} err="failed to get container status \"010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413\": rpc error: code = NotFound desc = could not find container \"010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413\": container with ID starting with 010a43f94ec035764b1710263443aecee202bf63ddd6f46924278d8065a8b413 not found: ID does not exist" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898006 4935 scope.go:117] "RemoveContainer" containerID="71a780f032841e0e4778281485674a6786913a58eb935c66f913a76d256ba68d" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898608 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-credential-keys\") pod \"6214d810-93b4-4537-a655-642df573048b\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898659 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-internal-tls-certs\") pod \"6214d810-93b4-4537-a655-642df573048b\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898681 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-public-tls-certs\") pod \"6214d810-93b4-4537-a655-642df573048b\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898712 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-config-data\") pod \"6214d810-93b4-4537-a655-642df573048b\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898801 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-combined-ca-bundle\") pod \"6214d810-93b4-4537-a655-642df573048b\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898827 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-scripts\") pod \"6214d810-93b4-4537-a655-642df573048b\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898849 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-fernet-keys\") pod \"6214d810-93b4-4537-a655-642df573048b\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.898880 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbsd8\" (UniqueName: \"kubernetes.io/projected/6214d810-93b4-4537-a655-642df573048b-kube-api-access-sbsd8\") pod \"6214d810-93b4-4537-a655-642df573048b\" (UID: \"6214d810-93b4-4537-a655-642df573048b\") " Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.900679 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.934322 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.935010 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6214d810-93b4-4537-a655-642df573048b" (UID: "6214d810-93b4-4537-a655-642df573048b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.935027 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6214d810-93b4-4537-a655-642df573048b-kube-api-access-sbsd8" (OuterVolumeSpecName: "kube-api-access-sbsd8") pod "6214d810-93b4-4537-a655-642df573048b" (UID: "6214d810-93b4-4537-a655-642df573048b"). InnerVolumeSpecName "kube-api-access-sbsd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.935119 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-scripts" (OuterVolumeSpecName: "scripts") pod "6214d810-93b4-4537-a655-642df573048b" (UID: "6214d810-93b4-4537-a655-642df573048b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.936691 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6214d810-93b4-4537-a655-642df573048b" (UID: "6214d810-93b4-4537-a655-642df573048b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.939234 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6214d810-93b4-4537-a655-642df573048b" (UID: "6214d810-93b4-4537-a655-642df573048b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.946792 4935 scope.go:117] "RemoveContainer" containerID="1c872576744b22c92353c71dde3535c32028ee33246e1225aa878fecf013d0b6" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.966997 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-config-data" (OuterVolumeSpecName: "config-data") pod "6214d810-93b4-4537-a655-642df573048b" (UID: "6214d810-93b4-4537-a655-642df573048b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.971058 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6214d810-93b4-4537-a655-642df573048b" (UID: "6214d810-93b4-4537-a655-642df573048b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:56 crc kubenswrapper[4935]: I1005 07:15:56.992114 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6214d810-93b4-4537-a655-642df573048b" (UID: "6214d810-93b4-4537-a655-642df573048b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.000925 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.000960 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.000969 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.000978 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.000986 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.000995 4935 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.001004 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbsd8\" (UniqueName: \"kubernetes.io/projected/6214d810-93b4-4537-a655-642df573048b-kube-api-access-sbsd8\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.001013 4935 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6214d810-93b4-4537-a655-642df573048b-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.067379 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.204110 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-79d9bc6587-tvhv2"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.204145 4935 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.204203 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data podName:53fe24e1-5efb-4422-ade0-16e315b7e328 nodeName:}" failed. No retries permitted until 2025-10-05 07:16:05.204189145 +0000 UTC m=+1399.086815595 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data") pod "rabbitmq-cell1-server-0" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.209371 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-79d9bc6587-tvhv2"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.253009 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.253372 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.253651 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.253704 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.254796 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.257547 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.266155 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.266239 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.548667 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-px6js" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerName="ovn-controller" probeResult="failure" output="" Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.557388 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936 is running failed: container process not found" containerID="eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.557816 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936 is running failed: container process not found" containerID="eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.558916 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936 is running failed: container process not found" containerID="eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.558951 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-px6js" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerName="ovn-controller" Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.559073 4935 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 05 07:15:57 crc kubenswrapper[4935]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-05T07:15:50Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 05 07:15:57 crc kubenswrapper[4935]: /etc/init.d/functions: line 589: 386 Alarm clock "$@" Oct 05 07:15:57 crc kubenswrapper[4935]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-px6js" message=< Oct 05 07:15:57 crc kubenswrapper[4935]: Exiting ovn-controller (1) [FAILED] Oct 05 07:15:57 crc kubenswrapper[4935]: Killing ovn-controller (1) [ OK ] Oct 05 07:15:57 crc kubenswrapper[4935]: Killing ovn-controller (1) with SIGKILL [ OK ] Oct 05 07:15:57 crc kubenswrapper[4935]: 2025-10-05T07:15:50Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 05 07:15:57 crc kubenswrapper[4935]: /etc/init.d/functions: line 589: 386 Alarm clock "$@" Oct 05 07:15:57 crc kubenswrapper[4935]: > Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.559128 4935 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 05 07:15:57 crc kubenswrapper[4935]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-05T07:15:50Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 05 07:15:57 crc kubenswrapper[4935]: /etc/init.d/functions: line 589: 386 Alarm clock "$@" Oct 05 07:15:57 crc kubenswrapper[4935]: > pod="openstack/ovn-controller-px6js" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerName="ovn-controller" containerID="cri-o://eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.559378 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-px6js" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerName="ovn-controller" containerID="cri-o://eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936" gracePeriod=22 Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.722276 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.751358 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.905313 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-px6js_971ba03f-bdf3-4788-a28d-76ae1c5c658c/ovn-controller/0.log" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.905787 4935 generic.go:334] "Generic (PLEG): container finished" podID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerID="eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936" exitCode=137 Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.905823 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-px6js" event={"ID":"971ba03f-bdf3-4788-a28d-76ae1c5c658c","Type":"ContainerDied","Data":"eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936"} Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.914496 4935 generic.go:334] "Generic (PLEG): container finished" podID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerID="9e3ecf50ee2e44f330efce8b076a27f4149d1c011a741f1c565a07c5137e9563" exitCode=0 Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.914572 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"53fe24e1-5efb-4422-ade0-16e315b7e328","Type":"ContainerDied","Data":"9e3ecf50ee2e44f330efce8b076a27f4149d1c011a741f1c565a07c5137e9563"} Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.916578 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bhxr\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-kube-api-access-2bhxr\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.916722 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.916793 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-plugins-conf\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.916828 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-erlang-cookie\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.916876 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8dcf22e2-2acd-4b49-a2bc-813a50f89409-erlang-cookie-secret\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.916921 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-confd\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.917038 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.917129 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8dcf22e2-2acd-4b49-a2bc-813a50f89409-pod-info\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.917253 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-plugins\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.917340 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-server-conf\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.917377 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-tls\") pod \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\" (UID: \"8dcf22e2-2acd-4b49-a2bc-813a50f89409\") " Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.919437 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.920295 4935 generic.go:334] "Generic (PLEG): container finished" podID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" containerID="cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a" exitCode=0 Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.920328 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dcf22e2-2acd-4b49-a2bc-813a50f89409","Type":"ContainerDied","Data":"cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a"} Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.920350 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dcf22e2-2acd-4b49-a2bc-813a50f89409","Type":"ContainerDied","Data":"923ad154d30fd0bd4a9a24e52741d6b23c1def8493cafe588f96b8b50acf6fda"} Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.920367 4935 scope.go:117] "RemoveContainer" containerID="cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.920376 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.922205 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.922335 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.923172 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.923913 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.924176 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dcf22e2-2acd-4b49-a2bc-813a50f89409-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.924948 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-kube-api-access-2bhxr" (OuterVolumeSpecName: "kube-api-access-2bhxr") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "kube-api-access-2bhxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.925215 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8dcf22e2-2acd-4b49-a2bc-813a50f89409-pod-info" (OuterVolumeSpecName: "pod-info") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.929025 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-px6js_971ba03f-bdf3-4788-a28d-76ae1c5c658c/ovn-controller/0.log" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.929080 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-px6js" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.939198 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data" (OuterVolumeSpecName: "config-data") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.956173 4935 scope.go:117] "RemoveContainer" containerID="7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.975649 4935 scope.go:117] "RemoveContainer" containerID="cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a" Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.976356 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a\": container with ID starting with cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a not found: ID does not exist" containerID="cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.976390 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a"} err="failed to get container status \"cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a\": rpc error: code = NotFound desc = could not find container \"cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a\": container with ID starting with cf2f68c65150649fdea1896c1dd3b9c97afe6328547eb9a7065c1004edc96e8a not found: ID does not exist" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.976415 4935 scope.go:117] "RemoveContainer" containerID="7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696" Oct 05 07:15:57 crc kubenswrapper[4935]: E1005 07:15:57.976982 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696\": container with ID starting with 7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696 not found: ID does not exist" containerID="7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.977014 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696"} err="failed to get container status \"7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696\": rpc error: code = NotFound desc = could not find container \"7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696\": container with ID starting with 7a66666d6a61f09e00d0c9542641746b184f6248be7d5251b8949bd53c939696 not found: ID does not exist" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.977299 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:15:57 crc kubenswrapper[4935]: I1005 07:15:57.982003 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-server-conf" (OuterVolumeSpecName: "server-conf") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018331 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-log-ovn\") pod \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018426 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-ovn-controller-tls-certs\") pod \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018486 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/971ba03f-bdf3-4788-a28d-76ae1c5c658c-scripts\") pod \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018595 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run-ovn\") pod \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018622 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-combined-ca-bundle\") pod \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018696 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7xds\" (UniqueName: \"kubernetes.io/projected/971ba03f-bdf3-4788-a28d-76ae1c5c658c-kube-api-access-b7xds\") pod \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018716 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run\") pod \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\" (UID: \"971ba03f-bdf3-4788-a28d-76ae1c5c658c\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018464 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "971ba03f-bdf3-4788-a28d-76ae1c5c658c" (UID: "971ba03f-bdf3-4788-a28d-76ae1c5c658c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.018821 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "971ba03f-bdf3-4788-a28d-76ae1c5c658c" (UID: "971ba03f-bdf3-4788-a28d-76ae1c5c658c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019048 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019063 4935 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8dcf22e2-2acd-4b49-a2bc-813a50f89409-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019074 4935 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019084 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019094 4935 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019102 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019111 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bhxr\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-kube-api-access-2bhxr\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019119 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019119 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run" (OuterVolumeSpecName: "var-run") pod "971ba03f-bdf3-4788-a28d-76ae1c5c658c" (UID: "971ba03f-bdf3-4788-a28d-76ae1c5c658c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019128 4935 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019172 4935 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8dcf22e2-2acd-4b49-a2bc-813a50f89409-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019187 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.019200 4935 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8dcf22e2-2acd-4b49-a2bc-813a50f89409-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.023765 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/971ba03f-bdf3-4788-a28d-76ae1c5c658c-scripts" (OuterVolumeSpecName: "scripts") pod "971ba03f-bdf3-4788-a28d-76ae1c5c658c" (UID: "971ba03f-bdf3-4788-a28d-76ae1c5c658c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.023995 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/971ba03f-bdf3-4788-a28d-76ae1c5c658c-kube-api-access-b7xds" (OuterVolumeSpecName: "kube-api-access-b7xds") pod "971ba03f-bdf3-4788-a28d-76ae1c5c658c" (UID: "971ba03f-bdf3-4788-a28d-76ae1c5c658c"). InnerVolumeSpecName "kube-api-access-b7xds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.031038 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8dcf22e2-2acd-4b49-a2bc-813a50f89409" (UID: "8dcf22e2-2acd-4b49-a2bc-813a50f89409"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.041941 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "971ba03f-bdf3-4788-a28d-76ae1c5c658c" (UID: "971ba03f-bdf3-4788-a28d-76ae1c5c658c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.048324 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.085015 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "971ba03f-bdf3-4788-a28d-76ae1c5c658c" (UID: "971ba03f-bdf3-4788-a28d-76ae1c5c658c"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.120659 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-server-conf\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.120750 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-confd\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.120796 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53fe24e1-5efb-4422-ade0-16e315b7e328-pod-info\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.120836 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-plugins\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.120867 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-erlang-cookie\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.120955 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.120972 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-plugins-conf\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121016 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53fe24e1-5efb-4422-ade0-16e315b7e328-erlang-cookie-secret\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121060 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9vhq\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-kube-api-access-v9vhq\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121076 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-tls\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121101 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"53fe24e1-5efb-4422-ade0-16e315b7e328\" (UID: \"53fe24e1-5efb-4422-ade0-16e315b7e328\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121370 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121389 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/971ba03f-bdf3-4788-a28d-76ae1c5c658c-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121401 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121412 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7xds\" (UniqueName: \"kubernetes.io/projected/971ba03f-bdf3-4788-a28d-76ae1c5c658c-kube-api-access-b7xds\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121421 4935 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/971ba03f-bdf3-4788-a28d-76ae1c5c658c-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121429 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/971ba03f-bdf3-4788-a28d-76ae1c5c658c-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.121438 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8dcf22e2-2acd-4b49-a2bc-813a50f89409-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.122121 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.122263 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.122762 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.123750 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/53fe24e1-5efb-4422-ade0-16e315b7e328-pod-info" (OuterVolumeSpecName: "pod-info") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.124085 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-kube-api-access-v9vhq" (OuterVolumeSpecName: "kube-api-access-v9vhq") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "kube-api-access-v9vhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.124265 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53fe24e1-5efb-4422-ade0-16e315b7e328-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.124267 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.127421 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.139374 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data" (OuterVolumeSpecName: "config-data") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.158036 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-server-conf" (OuterVolumeSpecName: "server-conf") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.213193 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "53fe24e1-5efb-4422-ade0-16e315b7e328" (UID: "53fe24e1-5efb-4422-ade0-16e315b7e328"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.225560 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.225610 4935 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.225626 4935 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53fe24e1-5efb-4422-ade0-16e315b7e328-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.225640 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9vhq\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-kube-api-access-v9vhq\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.225654 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.225701 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.225715 4935 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53fe24e1-5efb-4422-ade0-16e315b7e328-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.226916 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.226950 4935 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53fe24e1-5efb-4422-ade0-16e315b7e328-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.226968 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.226986 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53fe24e1-5efb-4422-ade0-16e315b7e328-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.243773 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.263671 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.269183 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.328669 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.793006 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6214d810-93b4-4537-a655-642df573048b" path="/var/lib/kubelet/pods/6214d810-93b4-4537-a655-642df573048b/volumes" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.794048 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" path="/var/lib/kubelet/pods/7c412f4e-54aa-4d6c-9606-753f97474c3b/volumes" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.794732 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" path="/var/lib/kubelet/pods/8dcf22e2-2acd-4b49-a2bc-813a50f89409/volumes" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.899457 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.932286 4935 generic.go:334] "Generic (PLEG): container finished" podID="c9b2e8af-97b0-4103-a241-ca473f299043" containerID="f1adbd1314cf52e09287126eae2dabba2ce549b2976316b1a565b87702fb1644" exitCode=0 Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.932345 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57fd77d677-fr4zj" event={"ID":"c9b2e8af-97b0-4103-a241-ca473f299043","Type":"ContainerDied","Data":"f1adbd1314cf52e09287126eae2dabba2ce549b2976316b1a565b87702fb1644"} Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.936133 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"53fe24e1-5efb-4422-ade0-16e315b7e328","Type":"ContainerDied","Data":"13bd2a8878775261a6323e48f9ceab16063a52e65164f2517b9a6afd89f9fdff"} Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.936165 4935 scope.go:117] "RemoveContainer" containerID="9e3ecf50ee2e44f330efce8b076a27f4149d1c011a741f1c565a07c5137e9563" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.936277 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.937349 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data-custom\") pod \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.937381 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data\") pod \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.937424 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkx65\" (UniqueName: \"kubernetes.io/projected/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-kube-api-access-qkx65\") pod \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.937485 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-combined-ca-bundle\") pod \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.937580 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-logs\") pod \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\" (UID: \"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b\") " Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.939954 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-logs" (OuterVolumeSpecName: "logs") pod "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" (UID: "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.943229 4935 generic.go:334] "Generic (PLEG): container finished" podID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerID="9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00" exitCode=0 Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.943294 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" event={"ID":"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b","Type":"ContainerDied","Data":"9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00"} Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.943322 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" event={"ID":"e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b","Type":"ContainerDied","Data":"38551ca6015da5ff6432e39dd2e14f955aeb6d46439d53443c31bfff7131960d"} Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.943371 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7cbc6494d6-8n275" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.943602 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-kube-api-access-qkx65" (OuterVolumeSpecName: "kube-api-access-qkx65") pod "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" (UID: "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b"). InnerVolumeSpecName "kube-api-access-qkx65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.944363 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" (UID: "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.946765 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.946869 4935 generic.go:334] "Generic (PLEG): container finished" podID="0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" containerID="24a61e5d809f2202e8aff2ffc919b4f809f3ea3e05132213aaafa01f0e1fc0d4" exitCode=0 Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.946939 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285","Type":"ContainerDied","Data":"24a61e5d809f2202e8aff2ffc919b4f809f3ea3e05132213aaafa01f0e1fc0d4"} Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.949740 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-px6js_971ba03f-bdf3-4788-a28d-76ae1c5c658c/ovn-controller/0.log" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.949775 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-px6js" event={"ID":"971ba03f-bdf3-4788-a28d-76ae1c5c658c","Type":"ContainerDied","Data":"5fcd8d9784ce4232eba94e1b2dafd37ef5063910d6bc79f00d465760e9da532e"} Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.949833 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-px6js" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.975161 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.984418 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.988746 4935 scope.go:117] "RemoveContainer" containerID="9c18cae5b21eea72b6ce894af8297def856aa4211aff70c789ca6375a34a3330" Oct 05 07:15:58 crc kubenswrapper[4935]: I1005 07:15:58.992684 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-px6js"] Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.003764 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-px6js"] Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.005814 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" (UID: "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.009284 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data" (OuterVolumeSpecName: "config-data") pod "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" (UID: "e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.023244 4935 scope.go:117] "RemoveContainer" containerID="9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.038983 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq8v4\" (UniqueName: \"kubernetes.io/projected/c9b2e8af-97b0-4103-a241-ca473f299043-kube-api-access-lq8v4\") pod \"c9b2e8af-97b0-4103-a241-ca473f299043\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039018 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-combined-ca-bundle\") pod \"c9b2e8af-97b0-4103-a241-ca473f299043\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039136 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9b2e8af-97b0-4103-a241-ca473f299043-logs\") pod \"c9b2e8af-97b0-4103-a241-ca473f299043\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039165 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data\") pod \"c9b2e8af-97b0-4103-a241-ca473f299043\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039224 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data-custom\") pod \"c9b2e8af-97b0-4103-a241-ca473f299043\" (UID: \"c9b2e8af-97b0-4103-a241-ca473f299043\") " Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039508 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039764 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039776 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039784 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039685 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9b2e8af-97b0-4103-a241-ca473f299043-logs" (OuterVolumeSpecName: "logs") pod "c9b2e8af-97b0-4103-a241-ca473f299043" (UID: "c9b2e8af-97b0-4103-a241-ca473f299043"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.039793 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkx65\" (UniqueName: \"kubernetes.io/projected/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b-kube-api-access-qkx65\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.042081 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c9b2e8af-97b0-4103-a241-ca473f299043" (UID: "c9b2e8af-97b0-4103-a241-ca473f299043"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.042786 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9b2e8af-97b0-4103-a241-ca473f299043-kube-api-access-lq8v4" (OuterVolumeSpecName: "kube-api-access-lq8v4") pod "c9b2e8af-97b0-4103-a241-ca473f299043" (UID: "c9b2e8af-97b0-4103-a241-ca473f299043"). InnerVolumeSpecName "kube-api-access-lq8v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.058276 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9b2e8af-97b0-4103-a241-ca473f299043" (UID: "c9b2e8af-97b0-4103-a241-ca473f299043"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.059065 4935 scope.go:117] "RemoveContainer" containerID="8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.086796 4935 scope.go:117] "RemoveContainer" containerID="9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.087237 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data" (OuterVolumeSpecName: "config-data") pod "c9b2e8af-97b0-4103-a241-ca473f299043" (UID: "c9b2e8af-97b0-4103-a241-ca473f299043"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: E1005 07:15:59.087745 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00\": container with ID starting with 9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00 not found: ID does not exist" containerID="9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.087793 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00"} err="failed to get container status \"9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00\": rpc error: code = NotFound desc = could not find container \"9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00\": container with ID starting with 9322c23a9563584f464e04b2b7fe88606ea90baf8b1d913ee0a3d9417e4bee00 not found: ID does not exist" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.087822 4935 scope.go:117] "RemoveContainer" containerID="8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a" Oct 05 07:15:59 crc kubenswrapper[4935]: E1005 07:15:59.088280 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a\": container with ID starting with 8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a not found: ID does not exist" containerID="8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.088317 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a"} err="failed to get container status \"8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a\": rpc error: code = NotFound desc = could not find container \"8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a\": container with ID starting with 8f8c7360c322185df555aa55f77e0956e399a71d461df1f5c2d4718c1e77be1a not found: ID does not exist" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.088345 4935 scope.go:117] "RemoveContainer" containerID="eaff04e8de22e9f2c82538718cdbee602767a65645b48058986d9cda28222936" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.102827 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.140356 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89kfs\" (UniqueName: \"kubernetes.io/projected/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-kube-api-access-89kfs\") pod \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.140416 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-config-data\") pod \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.140456 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-combined-ca-bundle\") pod \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\" (UID: \"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285\") " Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.140696 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9b2e8af-97b0-4103-a241-ca473f299043-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.140712 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.140722 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.140732 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq8v4\" (UniqueName: \"kubernetes.io/projected/c9b2e8af-97b0-4103-a241-ca473f299043-kube-api-access-lq8v4\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.140740 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9b2e8af-97b0-4103-a241-ca473f299043-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.143511 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-kube-api-access-89kfs" (OuterVolumeSpecName: "kube-api-access-89kfs") pod "0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" (UID: "0f2e339d-5a3f-4ab4-91a3-3fc73a96f285"). InnerVolumeSpecName "kube-api-access-89kfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.157652 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" (UID: "0f2e339d-5a3f-4ab4-91a3-3fc73a96f285"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.159373 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-config-data" (OuterVolumeSpecName: "config-data") pod "0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" (UID: "0f2e339d-5a3f-4ab4-91a3-3fc73a96f285"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.241802 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89kfs\" (UniqueName: \"kubernetes.io/projected/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-kube-api-access-89kfs\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.241844 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.241854 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.275921 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7cbc6494d6-8n275"] Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.276751 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7cbc6494d6-8n275"] Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.486382 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-74989f54d8-bw287" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.486526 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-74989f54d8-bw287" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.961129 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57fd77d677-fr4zj" event={"ID":"c9b2e8af-97b0-4103-a241-ca473f299043","Type":"ContainerDied","Data":"85dd90408525c28687ced0ae2ac7d321bf695ebcedcee982cb394973277cd908"} Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.961160 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57fd77d677-fr4zj" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.961195 4935 scope.go:117] "RemoveContainer" containerID="f1adbd1314cf52e09287126eae2dabba2ce549b2976316b1a565b87702fb1644" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.965207 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0f2e339d-5a3f-4ab4-91a3-3fc73a96f285","Type":"ContainerDied","Data":"080679fceb913252b8ca0d46672b01abb1ff18b9b01f1e11fa93f71b02709f61"} Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.965242 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 07:15:59 crc kubenswrapper[4935]: I1005 07:15:59.984852 4935 scope.go:117] "RemoveContainer" containerID="a5103fe4899d57404d97b7ee8251771675c6a622925145cac9caa880ee1b5c91" Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.013420 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-57fd77d677-fr4zj"] Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.025748 4935 scope.go:117] "RemoveContainer" containerID="24a61e5d809f2202e8aff2ffc919b4f809f3ea3e05132213aaafa01f0e1fc0d4" Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.026866 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-57fd77d677-fr4zj"] Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.030758 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.036728 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.785702 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" path="/var/lib/kubelet/pods/0f2e339d-5a3f-4ab4-91a3-3fc73a96f285/volumes" Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.786655 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53fe24e1-5efb-4422-ade0-16e315b7e328" path="/var/lib/kubelet/pods/53fe24e1-5efb-4422-ade0-16e315b7e328/volumes" Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.787207 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" path="/var/lib/kubelet/pods/971ba03f-bdf3-4788-a28d-76ae1c5c658c/volumes" Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.788266 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" path="/var/lib/kubelet/pods/c9b2e8af-97b0-4103-a241-ca473f299043/volumes" Oct 05 07:16:00 crc kubenswrapper[4935]: I1005 07:16:00.788787 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" path="/var/lib/kubelet/pods/e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b/volumes" Oct 05 07:16:02 crc kubenswrapper[4935]: E1005 07:16:02.253248 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:02 crc kubenswrapper[4935]: E1005 07:16:02.254472 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:02 crc kubenswrapper[4935]: E1005 07:16:02.254886 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:02 crc kubenswrapper[4935]: E1005 07:16:02.254983 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:16:02 crc kubenswrapper[4935]: E1005 07:16:02.255382 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:02 crc kubenswrapper[4935]: E1005 07:16:02.257186 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:02 crc kubenswrapper[4935]: E1005 07:16:02.258405 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:02 crc kubenswrapper[4935]: E1005 07:16:02.258467 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178046 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-66dvp"] Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178520 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerName="ovn-controller" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178547 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerName="ovn-controller" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178573 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerName="cinder-scheduler" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178586 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerName="cinder-scheduler" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178609 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77853c76-2bbe-42e5-a0ef-03d5c5fe2402" containerName="nova-cell1-conductor-conductor" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178622 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="77853c76-2bbe-42e5-a0ef-03d5c5fe2402" containerName="nova-cell1-conductor-conductor" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178640 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bedc55a-2396-4249-ac54-a753a67dca72" containerName="galera" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178652 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bedc55a-2396-4249-ac54-a753a67dca72" containerName="galera" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178673 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="proxy-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178687 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="proxy-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178701 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerName="glance-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178713 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerName="glance-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178732 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178744 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178759 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" containerName="barbican-worker" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178772 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" containerName="barbican-worker" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178791 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="ceilometer-notification-agent" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178804 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="ceilometer-notification-agent" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178821 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerName="setup-container" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178834 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerName="setup-container" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178861 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerName="glance-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178872 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerName="glance-log" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178886 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178920 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178935 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178948 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.178973 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" containerName="glance-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.178985 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" containerName="glance-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179002 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="sg-core" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179017 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="sg-core" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179036 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0095740d-765a-4cba-b0c8-d36592fad800" containerName="mariadb-account-delete" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179048 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0095740d-765a-4cba-b0c8-d36592fad800" containerName="mariadb-account-delete" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179066 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerName="barbican-keystone-listener" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179078 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerName="barbican-keystone-listener" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179094 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" containerName="nova-cell0-conductor-conductor" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179106 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" containerName="nova-cell0-conductor-conductor" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179136 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="ovsdbserver-nb" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179151 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="ovsdbserver-nb" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179172 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerName="dnsmasq-dns" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179187 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerName="dnsmasq-dns" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179203 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6214d810-93b4-4537-a655-642df573048b" containerName="keystone-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179219 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6214d810-93b4-4537-a655-642df573048b" containerName="keystone-api" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179243 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" containerName="mysql-bootstrap" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179255 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" containerName="mysql-bootstrap" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179271 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179283 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179299 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" containerName="rabbitmq" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179312 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" containerName="rabbitmq" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179329 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-metadata" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179341 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-metadata" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179354 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" containerName="galera" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179537 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" containerName="galera" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179567 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerName="placement-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179579 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerName="placement-log" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179599 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179611 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179630 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerName="probe" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179642 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerName="probe" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179662 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" containerName="glance-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179674 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" containerName="glance-log" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179690 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="ovn-northd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179702 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="ovn-northd" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179724 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="ceilometer-central-agent" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179736 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="ceilometer-central-agent" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179749 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerName="rabbitmq" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179761 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerName="rabbitmq" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179779 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bedc55a-2396-4249-ac54-a753a67dca72" containerName="mysql-bootstrap" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179792 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bedc55a-2396-4249-ac54-a753a67dca72" containerName="mysql-bootstrap" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179814 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179827 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179848 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerName="init" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179860 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerName="init" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179877 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b82213d6-93fa-48fc-808b-36f902cce15c" containerName="memcached" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179911 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b82213d6-93fa-48fc-808b-36f902cce15c" containerName="memcached" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179937 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerName="placement-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179949 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerName="placement-api" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179969 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.179981 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-api" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.179998 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" containerName="setup-container" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180010 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" containerName="setup-container" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180026 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerName="proxy-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180038 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerName="proxy-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180056 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerName="barbican-keystone-listener-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180068 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerName="barbican-keystone-listener-log" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180196 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerName="proxy-server" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180211 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerName="proxy-server" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180236 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" containerName="nova-scheduler-scheduler" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180248 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" containerName="nova-scheduler-scheduler" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180267 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180279 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180298 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerName="ovsdbserver-sb" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180310 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerName="ovsdbserver-sb" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180329 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f950314f-27aa-4a44-b13a-4b4f3a7495ab" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180341 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f950314f-27aa-4a44-b13a-4b4f3a7495ab" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180356 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180368 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-log" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180387 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca591a2f-54ed-41b1-aa10-be879b218844" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180399 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca591a2f-54ed-41b1-aa10-be879b218844" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180418 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c710b93b-c2fe-4a03-ba8c-f280a9d67da6" containerName="kube-state-metrics" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180431 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c710b93b-c2fe-4a03-ba8c-f280a9d67da6" containerName="kube-state-metrics" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180449 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ccd5ed-23d9-4d0b-99fb-68655c0199b5" containerName="mariadb-account-delete" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180461 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ccd5ed-23d9-4d0b-99fb-68655c0199b5" containerName="mariadb-account-delete" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180479 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180491 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: E1005 07:16:03.180505 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" containerName="barbican-worker-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180516 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" containerName="barbican-worker-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180813 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="971ba03f-bdf3-4788-a28d-76ae1c5c658c" containerName="ovn-controller" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180846 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180868 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerName="placement-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180887 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0095740d-765a-4cba-b0c8-d36592fad800" containerName="mariadb-account-delete" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180938 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180956 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f27b0edc-d2a4-418e-8c62-82f1287cc1b4" containerName="ovsdbserver-nb" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180968 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerName="glance-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.180987 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6ccd5ed-23d9-4d0b-99fb-68655c0199b5" containerName="mariadb-account-delete" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181009 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="af041b3a-6ed3-4c8d-836b-a9b69f95360b" containerName="galera" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181022 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181035 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="53fe24e1-5efb-4422-ade0-16e315b7e328" containerName="rabbitmq" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181053 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerName="probe" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181069 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181090 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a7ceab-efbb-4ba4-a5f2-c6702abbc314" containerName="placement-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181112 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerName="ovsdbserver-sb" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181125 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c422d4a-a835-443a-9c6d-3161bc80c1aa" containerName="barbican-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181139 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" containerName="glance-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181157 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" containerName="barbican-worker-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181179 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="ceilometer-central-agent" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181197 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="77853c76-2bbe-42e5-a0ef-03d5c5fe2402" containerName="nova-cell1-conductor-conductor" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181211 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c710b93b-c2fe-4a03-ba8c-f280a9d67da6" containerName="kube-state-metrics" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181227 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" containerName="dnsmasq-dns" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181247 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9acce3fe-5a42-49e7-b3e7-a72f01af5736" containerName="cinder-scheduler" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181270 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2e339d-5a3f-4ab4-91a3-3fc73a96f285" containerName="nova-cell0-conductor-conductor" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181299 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="74e20071-1ff2-4fae-ae8c-beb9dbce35ce" containerName="glance-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181326 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b82213d6-93fa-48fc-808b-36f902cce15c" containerName="memcached" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181348 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerName="barbican-keystone-listener" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181372 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5efc4a0-badc-4f8c-a58b-ef0f5e42f93b" containerName="barbican-keystone-listener-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181396 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c7acc7a-797f-46cd-b298-1f28fd2951d6" containerName="cinder-api-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181424 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f950314f-27aa-4a44-b13a-4b4f3a7495ab" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181441 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6d217a2-8ab3-46a8-a70b-58cb5d5cad32" containerName="nova-scheduler-scheduler" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181461 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b2e8af-97b0-4103-a241-ca473f299043" containerName="barbican-worker" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181486 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-metadata" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181502 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="sg-core" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181524 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerName="proxy-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181541 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c512e7c-9418-43b4-8235-4d62fa487556" containerName="glance-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181561 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bedc55a-2396-4249-ac54-a753a67dca72" containerName="galera" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181581 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca591a2f-54ed-41b1-aa10-be879b218844" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181603 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d984f8-20c4-4b30-b767-c6106aa279cf" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181616 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6214d810-93b4-4537-a655-642df573048b" containerName="keystone-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181636 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="openstack-network-exporter" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181649 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f1eb003-e111-4a86-b6c7-a899aa0426a0" containerName="nova-metadata-log" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181669 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="ceilometer-notification-agent" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181688 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba39a0b-f275-4075-9419-a354a3c5fa5e" containerName="proxy-httpd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181711 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcf22e2-2acd-4b49-a2bc-813a50f89409" containerName="rabbitmq" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181728 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e99221-24f0-4cbe-bb3c-4e1e70acd4ba" containerName="proxy-server" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181743 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c412f4e-54aa-4d6c-9606-753f97474c3b" containerName="ovn-northd" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.181760 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6a2e9cc-e6b4-4af2-bc01-0e8b4038db9c" containerName="nova-api-api" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.183610 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.194156 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66dvp"] Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.304730 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-catalog-content\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.304806 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qddrm\" (UniqueName: \"kubernetes.io/projected/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-kube-api-access-qddrm\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.304844 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-utilities\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.406667 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-catalog-content\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.406749 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qddrm\" (UniqueName: \"kubernetes.io/projected/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-kube-api-access-qddrm\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.406798 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-utilities\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.407222 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-catalog-content\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.407266 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-utilities\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.434239 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qddrm\" (UniqueName: \"kubernetes.io/projected/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-kube-api-access-qddrm\") pod \"redhat-operators-66dvp\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.501484 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:03 crc kubenswrapper[4935]: I1005 07:16:03.956565 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66dvp"] Oct 05 07:16:04 crc kubenswrapper[4935]: I1005 07:16:04.005967 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66dvp" event={"ID":"66b9e1fa-5995-4685-9d9c-e38469c4e7a1","Type":"ContainerStarted","Data":"b5f56267a7b38e57001f64e6ad307004df524ca881c0389b2cbb5796cfeb6ed0"} Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.017737 4935 generic.go:334] "Generic (PLEG): container finished" podID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerID="cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6" exitCode=0 Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.017784 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66dvp" event={"ID":"66b9e1fa-5995-4685-9d9c-e38469c4e7a1","Type":"ContainerDied","Data":"cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6"} Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.021818 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.898268 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.944772 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-internal-tls-certs\") pod \"27a28306-70be-4556-9659-64999f775195\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.944823 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-combined-ca-bundle\") pod \"27a28306-70be-4556-9659-64999f775195\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.944842 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-ovndb-tls-certs\") pod \"27a28306-70be-4556-9659-64999f775195\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.944883 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-public-tls-certs\") pod \"27a28306-70be-4556-9659-64999f775195\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.944964 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7h4c\" (UniqueName: \"kubernetes.io/projected/27a28306-70be-4556-9659-64999f775195-kube-api-access-s7h4c\") pod \"27a28306-70be-4556-9659-64999f775195\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.944994 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-config\") pod \"27a28306-70be-4556-9659-64999f775195\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.945013 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-httpd-config\") pod \"27a28306-70be-4556-9659-64999f775195\" (UID: \"27a28306-70be-4556-9659-64999f775195\") " Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.964843 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "27a28306-70be-4556-9659-64999f775195" (UID: "27a28306-70be-4556-9659-64999f775195"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.968171 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27a28306-70be-4556-9659-64999f775195-kube-api-access-s7h4c" (OuterVolumeSpecName: "kube-api-access-s7h4c") pod "27a28306-70be-4556-9659-64999f775195" (UID: "27a28306-70be-4556-9659-64999f775195"). InnerVolumeSpecName "kube-api-access-s7h4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.991949 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27a28306-70be-4556-9659-64999f775195" (UID: "27a28306-70be-4556-9659-64999f775195"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:16:05 crc kubenswrapper[4935]: I1005 07:16:05.995580 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "27a28306-70be-4556-9659-64999f775195" (UID: "27a28306-70be-4556-9659-64999f775195"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.006454 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-config" (OuterVolumeSpecName: "config") pod "27a28306-70be-4556-9659-64999f775195" (UID: "27a28306-70be-4556-9659-64999f775195"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.019293 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "27a28306-70be-4556-9659-64999f775195" (UID: "27a28306-70be-4556-9659-64999f775195"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.027905 4935 generic.go:334] "Generic (PLEG): container finished" podID="27a28306-70be-4556-9659-64999f775195" containerID="a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10" exitCode=0 Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.027966 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56bb7fd5-dmzld" event={"ID":"27a28306-70be-4556-9659-64999f775195","Type":"ContainerDied","Data":"a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10"} Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.027978 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56bb7fd5-dmzld" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.027994 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56bb7fd5-dmzld" event={"ID":"27a28306-70be-4556-9659-64999f775195","Type":"ContainerDied","Data":"cda4dff46d5e8b986aec637f976286eebb3d0f536a6e9ad90c6af7da74c71303"} Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.028008 4935 scope.go:117] "RemoveContainer" containerID="16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.031098 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66dvp" event={"ID":"66b9e1fa-5995-4685-9d9c-e38469c4e7a1","Type":"ContainerStarted","Data":"fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047"} Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.045052 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "27a28306-70be-4556-9659-64999f775195" (UID: "27a28306-70be-4556-9659-64999f775195"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.048585 4935 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.048698 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7h4c\" (UniqueName: \"kubernetes.io/projected/27a28306-70be-4556-9659-64999f775195-kube-api-access-s7h4c\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.048774 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.048836 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.048932 4935 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.049002 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.049061 4935 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a28306-70be-4556-9659-64999f775195-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.055625 4935 scope.go:117] "RemoveContainer" containerID="a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.072998 4935 scope.go:117] "RemoveContainer" containerID="16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000" Oct 05 07:16:06 crc kubenswrapper[4935]: E1005 07:16:06.073417 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000\": container with ID starting with 16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000 not found: ID does not exist" containerID="16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.073528 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000"} err="failed to get container status \"16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000\": rpc error: code = NotFound desc = could not find container \"16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000\": container with ID starting with 16f97a03fa3cec9b8e14f794543bf39a91c7dfe33659e7c9a9d650e1bcdec000 not found: ID does not exist" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.073606 4935 scope.go:117] "RemoveContainer" containerID="a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10" Oct 05 07:16:06 crc kubenswrapper[4935]: E1005 07:16:06.074015 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10\": container with ID starting with a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10 not found: ID does not exist" containerID="a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.074046 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10"} err="failed to get container status \"a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10\": rpc error: code = NotFound desc = could not find container \"a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10\": container with ID starting with a119d3dfd93cfafe1745081bb304344984169878463350152d43b952c6149e10 not found: ID does not exist" Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.434318 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56bb7fd5-dmzld"] Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.445365 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-56bb7fd5-dmzld"] Oct 05 07:16:06 crc kubenswrapper[4935]: I1005 07:16:06.796489 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27a28306-70be-4556-9659-64999f775195" path="/var/lib/kubelet/pods/27a28306-70be-4556-9659-64999f775195/volumes" Oct 05 07:16:07 crc kubenswrapper[4935]: I1005 07:16:07.043405 4935 generic.go:334] "Generic (PLEG): container finished" podID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerID="fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047" exitCode=0 Oct 05 07:16:07 crc kubenswrapper[4935]: I1005 07:16:07.043520 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66dvp" event={"ID":"66b9e1fa-5995-4685-9d9c-e38469c4e7a1","Type":"ContainerDied","Data":"fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047"} Oct 05 07:16:07 crc kubenswrapper[4935]: E1005 07:16:07.253912 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:07 crc kubenswrapper[4935]: E1005 07:16:07.254578 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:07 crc kubenswrapper[4935]: E1005 07:16:07.255279 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:07 crc kubenswrapper[4935]: E1005 07:16:07.255316 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:16:07 crc kubenswrapper[4935]: E1005 07:16:07.255597 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:07 crc kubenswrapper[4935]: E1005 07:16:07.257450 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:07 crc kubenswrapper[4935]: E1005 07:16:07.259639 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:07 crc kubenswrapper[4935]: E1005 07:16:07.259695 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:16:08 crc kubenswrapper[4935]: I1005 07:16:08.057098 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66dvp" event={"ID":"66b9e1fa-5995-4685-9d9c-e38469c4e7a1","Type":"ContainerStarted","Data":"657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5"} Oct 05 07:16:08 crc kubenswrapper[4935]: I1005 07:16:08.077618 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-66dvp" podStartSLOduration=2.624499236 podStartE2EDuration="5.077597049s" podCreationTimestamp="2025-10-05 07:16:03 +0000 UTC" firstStartedPulling="2025-10-05 07:16:05.021552263 +0000 UTC m=+1398.904178723" lastFinishedPulling="2025-10-05 07:16:07.474650086 +0000 UTC m=+1401.357276536" observedRunningTime="2025-10-05 07:16:08.072636448 +0000 UTC m=+1401.955262918" watchObservedRunningTime="2025-10-05 07:16:08.077597049 +0000 UTC m=+1401.960223509" Oct 05 07:16:12 crc kubenswrapper[4935]: E1005 07:16:12.253570 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:12 crc kubenswrapper[4935]: E1005 07:16:12.254596 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:12 crc kubenswrapper[4935]: E1005 07:16:12.254821 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:12 crc kubenswrapper[4935]: E1005 07:16:12.254954 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:12 crc kubenswrapper[4935]: E1005 07:16:12.254990 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:16:12 crc kubenswrapper[4935]: E1005 07:16:12.256344 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:12 crc kubenswrapper[4935]: E1005 07:16:12.257561 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:12 crc kubenswrapper[4935]: E1005 07:16:12.257609 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:16:13 crc kubenswrapper[4935]: I1005 07:16:13.502295 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:13 crc kubenswrapper[4935]: I1005 07:16:13.502750 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:13 crc kubenswrapper[4935]: I1005 07:16:13.573232 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:14 crc kubenswrapper[4935]: I1005 07:16:14.157942 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:14 crc kubenswrapper[4935]: I1005 07:16:14.217387 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66dvp"] Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.134684 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-66dvp" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerName="registry-server" containerID="cri-o://657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5" gracePeriod=2 Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.633151 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.725016 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-utilities\") pod \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.725115 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-catalog-content\") pod \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.725166 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qddrm\" (UniqueName: \"kubernetes.io/projected/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-kube-api-access-qddrm\") pod \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\" (UID: \"66b9e1fa-5995-4685-9d9c-e38469c4e7a1\") " Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.726741 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-utilities" (OuterVolumeSpecName: "utilities") pod "66b9e1fa-5995-4685-9d9c-e38469c4e7a1" (UID: "66b9e1fa-5995-4685-9d9c-e38469c4e7a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.732116 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-kube-api-access-qddrm" (OuterVolumeSpecName: "kube-api-access-qddrm") pod "66b9e1fa-5995-4685-9d9c-e38469c4e7a1" (UID: "66b9e1fa-5995-4685-9d9c-e38469c4e7a1"). InnerVolumeSpecName "kube-api-access-qddrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.826641 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qddrm\" (UniqueName: \"kubernetes.io/projected/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-kube-api-access-qddrm\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:16 crc kubenswrapper[4935]: I1005 07:16:16.826966 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.140618 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66b9e1fa-5995-4685-9d9c-e38469c4e7a1" (UID: "66b9e1fa-5995-4685-9d9c-e38469c4e7a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.145335 4935 generic.go:334] "Generic (PLEG): container finished" podID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerID="657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5" exitCode=0 Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.145378 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66dvp" event={"ID":"66b9e1fa-5995-4685-9d9c-e38469c4e7a1","Type":"ContainerDied","Data":"657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5"} Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.145403 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66dvp" event={"ID":"66b9e1fa-5995-4685-9d9c-e38469c4e7a1","Type":"ContainerDied","Data":"b5f56267a7b38e57001f64e6ad307004df524ca881c0389b2cbb5796cfeb6ed0"} Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.145429 4935 scope.go:117] "RemoveContainer" containerID="657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.145430 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66dvp" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.166464 4935 scope.go:117] "RemoveContainer" containerID="fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.178990 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66dvp"] Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.182124 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-66dvp"] Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.220293 4935 scope.go:117] "RemoveContainer" containerID="cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.231271 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b9e1fa-5995-4685-9d9c-e38469c4e7a1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.245398 4935 scope.go:117] "RemoveContainer" containerID="657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5" Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.245849 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5\": container with ID starting with 657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5 not found: ID does not exist" containerID="657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.245907 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5"} err="failed to get container status \"657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5\": rpc error: code = NotFound desc = could not find container \"657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5\": container with ID starting with 657de661925a567b9de70d053718dd15fcaccb4eb264cb877a7275305c3ad7b5 not found: ID does not exist" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.245934 4935 scope.go:117] "RemoveContainer" containerID="fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047" Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.246190 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047\": container with ID starting with fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047 not found: ID does not exist" containerID="fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.246218 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047"} err="failed to get container status \"fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047\": rpc error: code = NotFound desc = could not find container \"fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047\": container with ID starting with fce2b695b5336d3e98d19a0c56e060710c692224f60c7236a9cd3326bece6047 not found: ID does not exist" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.246234 4935 scope.go:117] "RemoveContainer" containerID="cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6" Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.246664 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6\": container with ID starting with cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6 not found: ID does not exist" containerID="cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6" Oct 05 07:16:17 crc kubenswrapper[4935]: I1005 07:16:17.246760 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6"} err="failed to get container status \"cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6\": rpc error: code = NotFound desc = could not find container \"cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6\": container with ID starting with cb55808d97fe45dc493a9f2817fdcce96a32e08a06771f3b2c15c59816f116d6 not found: ID does not exist" Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.253095 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.253510 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.253810 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.253889 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.254384 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.255702 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.256707 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:16:17 crc kubenswrapper[4935]: E1005 07:16:17.256734 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-pl4nq" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:16:18 crc kubenswrapper[4935]: I1005 07:16:18.792325 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" path="/var/lib/kubelet/pods/66b9e1fa-5995-4685-9d9c-e38469c4e7a1/volumes" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.858091 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pl4nq_aac8194b-9784-4eed-8c35-fb65a584b525/ovs-vswitchd/0.log" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.859163 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.973526 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcrnr\" (UniqueName: \"kubernetes.io/projected/aac8194b-9784-4eed-8c35-fb65a584b525-kube-api-access-xcrnr\") pod \"aac8194b-9784-4eed-8c35-fb65a584b525\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.973629 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-log\") pod \"aac8194b-9784-4eed-8c35-fb65a584b525\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.973660 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-lib\") pod \"aac8194b-9784-4eed-8c35-fb65a584b525\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.973882 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-run\") pod \"aac8194b-9784-4eed-8c35-fb65a584b525\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.973937 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-etc-ovs\") pod \"aac8194b-9784-4eed-8c35-fb65a584b525\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.973925 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-log" (OuterVolumeSpecName: "var-log") pod "aac8194b-9784-4eed-8c35-fb65a584b525" (UID: "aac8194b-9784-4eed-8c35-fb65a584b525"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.973956 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-lib" (OuterVolumeSpecName: "var-lib") pod "aac8194b-9784-4eed-8c35-fb65a584b525" (UID: "aac8194b-9784-4eed-8c35-fb65a584b525"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.973975 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac8194b-9784-4eed-8c35-fb65a584b525-scripts\") pod \"aac8194b-9784-4eed-8c35-fb65a584b525\" (UID: \"aac8194b-9784-4eed-8c35-fb65a584b525\") " Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.974004 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-run" (OuterVolumeSpecName: "var-run") pod "aac8194b-9784-4eed-8c35-fb65a584b525" (UID: "aac8194b-9784-4eed-8c35-fb65a584b525"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.974000 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "aac8194b-9784-4eed-8c35-fb65a584b525" (UID: "aac8194b-9784-4eed-8c35-fb65a584b525"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.974736 4935 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-log\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.974763 4935 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-lib\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.974781 4935 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.974796 4935 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/aac8194b-9784-4eed-8c35-fb65a584b525-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.975394 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aac8194b-9784-4eed-8c35-fb65a584b525-scripts" (OuterVolumeSpecName: "scripts") pod "aac8194b-9784-4eed-8c35-fb65a584b525" (UID: "aac8194b-9784-4eed-8c35-fb65a584b525"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:16:19 crc kubenswrapper[4935]: I1005 07:16:19.981042 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac8194b-9784-4eed-8c35-fb65a584b525-kube-api-access-xcrnr" (OuterVolumeSpecName: "kube-api-access-xcrnr") pod "aac8194b-9784-4eed-8c35-fb65a584b525" (UID: "aac8194b-9784-4eed-8c35-fb65a584b525"). InnerVolumeSpecName "kube-api-access-xcrnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.076451 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcrnr\" (UniqueName: \"kubernetes.io/projected/aac8194b-9784-4eed-8c35-fb65a584b525-kube-api-access-xcrnr\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.076487 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aac8194b-9784-4eed-8c35-fb65a584b525-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.188041 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pl4nq_aac8194b-9784-4eed-8c35-fb65a584b525/ovs-vswitchd/0.log" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.189696 4935 generic.go:334] "Generic (PLEG): container finished" podID="aac8194b-9784-4eed-8c35-fb65a584b525" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" exitCode=137 Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.189814 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pl4nq" event={"ID":"aac8194b-9784-4eed-8c35-fb65a584b525","Type":"ContainerDied","Data":"ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8"} Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.189870 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pl4nq" event={"ID":"aac8194b-9784-4eed-8c35-fb65a584b525","Type":"ContainerDied","Data":"3a134aaab30465b85e8f3f839977f757b39b55e191ba8e33e47ccc0612c14aea"} Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.189950 4935 scope.go:117] "RemoveContainer" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.190185 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pl4nq" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.210992 4935 generic.go:334] "Generic (PLEG): container finished" podID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerID="fbfc0f5f81ecb328f21528e6e7fd64b97e302cbe3748716791576b9c52740ca0" exitCode=137 Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.211032 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"fbfc0f5f81ecb328f21528e6e7fd64b97e302cbe3748716791576b9c52740ca0"} Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.229085 4935 scope.go:117] "RemoveContainer" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.283609 4935 scope.go:117] "RemoveContainer" containerID="2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.301314 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-pl4nq"] Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.307634 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-pl4nq"] Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.316980 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.317000 4935 scope.go:117] "RemoveContainer" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" Oct 05 07:16:20 crc kubenswrapper[4935]: E1005 07:16:20.317581 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8\": container with ID starting with ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8 not found: ID does not exist" containerID="ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.317619 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8"} err="failed to get container status \"ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8\": rpc error: code = NotFound desc = could not find container \"ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8\": container with ID starting with ea5ba058d9df08c5b4c658e28ada42ae88802468d4f7a6fdb22e65cda04682e8 not found: ID does not exist" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.317646 4935 scope.go:117] "RemoveContainer" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" Oct 05 07:16:20 crc kubenswrapper[4935]: E1005 07:16:20.318032 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e\": container with ID starting with 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e not found: ID does not exist" containerID="81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.318054 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e"} err="failed to get container status \"81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e\": rpc error: code = NotFound desc = could not find container \"81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e\": container with ID starting with 81271c559a5013c86b93b03fed1aea7356895a0ce0f24f7549f4b5c493db4f8e not found: ID does not exist" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.318069 4935 scope.go:117] "RemoveContainer" containerID="2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095" Oct 05 07:16:20 crc kubenswrapper[4935]: E1005 07:16:20.318287 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095\": container with ID starting with 2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095 not found: ID does not exist" containerID="2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.318310 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095"} err="failed to get container status \"2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095\": rpc error: code = NotFound desc = could not find container \"2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095\": container with ID starting with 2d0aeb791d0091d1e17133bf8af6cc4c06d8df2441478dd54ee2162ac0b72095 not found: ID does not exist" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.482809 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") pod \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.482884 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.482984 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr2vk\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-kube-api-access-gr2vk\") pod \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.483080 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-lock\") pod \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.483177 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-cache\") pod \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\" (UID: \"10a5c868-ffd7-44de-8f47-feb4c0ce9121\") " Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.483578 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-lock" (OuterVolumeSpecName: "lock") pod "10a5c868-ffd7-44de-8f47-feb4c0ce9121" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.483812 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-cache" (OuterVolumeSpecName: "cache") pod "10a5c868-ffd7-44de-8f47-feb4c0ce9121" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.487360 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "swift") pod "10a5c868-ffd7-44de-8f47-feb4c0ce9121" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.487688 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "10a5c868-ffd7-44de-8f47-feb4c0ce9121" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.488336 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-kube-api-access-gr2vk" (OuterVolumeSpecName: "kube-api-access-gr2vk") pod "10a5c868-ffd7-44de-8f47-feb4c0ce9121" (UID: "10a5c868-ffd7-44de-8f47-feb4c0ce9121"). InnerVolumeSpecName "kube-api-access-gr2vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.585673 4935 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-lock\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.585716 4935 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/10a5c868-ffd7-44de-8f47-feb4c0ce9121-cache\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.585728 4935 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.585769 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.585782 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr2vk\" (UniqueName: \"kubernetes.io/projected/10a5c868-ffd7-44de-8f47-feb4c0ce9121-kube-api-access-gr2vk\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.600044 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.686819 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:20 crc kubenswrapper[4935]: I1005 07:16:20.788388 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" path="/var/lib/kubelet/pods/aac8194b-9784-4eed-8c35-fb65a584b525/volumes" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.235775 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"10a5c868-ffd7-44de-8f47-feb4c0ce9121","Type":"ContainerDied","Data":"58c7a776ac2b258d5904745678ab9ccdba0ca364da2e464d53151208cc40c811"} Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.235872 4935 scope.go:117] "RemoveContainer" containerID="fbfc0f5f81ecb328f21528e6e7fd64b97e302cbe3748716791576b9c52740ca0" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.236008 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.278674 4935 scope.go:117] "RemoveContainer" containerID="f8b9ec64a993b5fc267da34859f562afb461e51a8c3449b8a4980e4dbec60aab" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.284506 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.296000 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.301408 4935 scope.go:117] "RemoveContainer" containerID="b90d88185b137f6f204daabf2f63d9c6ca4190f72a5e20e02ad439fe47b1d9e2" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.323944 4935 scope.go:117] "RemoveContainer" containerID="8ae6a7f5c8eb347b93d521a8aa9df7e0b4edee491ce3dd6e4dd29fd1cdc12733" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.346186 4935 scope.go:117] "RemoveContainer" containerID="5811d6a7453e1bdaa6a37fe17b4607303b6b13e3e3d1fe5e319d227f3a5e3efb" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.364240 4935 scope.go:117] "RemoveContainer" containerID="f07908304e5257af7f03f59b8ba6637156f346dca17601807d91da7efcc445e9" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.381758 4935 scope.go:117] "RemoveContainer" containerID="06e262448da43be1f59f8206d3ca149b097157a5612b8e5d368736aac9a62a57" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.401172 4935 scope.go:117] "RemoveContainer" containerID="ba8102ea51bc30c759201a866a8431d1baaf2f6f474cc96b36529ea5a04451ec" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.416485 4935 scope.go:117] "RemoveContainer" containerID="cd237a014093f689d7df188672e570ebe94fb2b0f49e1d1a6c58473d64c3b740" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.431091 4935 scope.go:117] "RemoveContainer" containerID="b18142ab07c004a98bff59b529be5e4c49e5ff54caf77526350e39ce8c20c39c" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.447456 4935 scope.go:117] "RemoveContainer" containerID="cd51894ecb6651dbb869d73d13d130a6f2fb7d28b0f7980f406450f51c4449b9" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.467608 4935 scope.go:117] "RemoveContainer" containerID="95b99cc19f9b362efc74b2612128d2a9a4247e4bfa40d9b7ca5e1c764c78c5c9" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.494677 4935 scope.go:117] "RemoveContainer" containerID="5fb020c3045a74d5048d67e9e8f939b0d2f9348bab133b9b65f5cfb8ff321504" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.512816 4935 scope.go:117] "RemoveContainer" containerID="22ba6a3f7d0c659cde8f9b4b899df8e9b5b426c7a3f7e76b86ffc508196c5584" Oct 05 07:16:21 crc kubenswrapper[4935]: I1005 07:16:21.534168 4935 scope.go:117] "RemoveContainer" containerID="a461381cdfc3c2eee74eb1bdcf7d5ebfdb0fb05a83a09c83b9bcd518fb51b2d5" Oct 05 07:16:22 crc kubenswrapper[4935]: I1005 07:16:22.646111 4935 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod8bf67c05-04fb-414a-aa6e-8cdc3831cb94"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod8bf67c05-04fb-414a-aa6e-8cdc3831cb94] : Timed out while waiting for systemd to remove kubepods-besteffort-pod8bf67c05_04fb_414a_aa6e_8cdc3831cb94.slice" Oct 05 07:16:22 crc kubenswrapper[4935]: E1005 07:16:22.646166 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod8bf67c05-04fb-414a-aa6e-8cdc3831cb94] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod8bf67c05-04fb-414a-aa6e-8cdc3831cb94] : Timed out while waiting for systemd to remove kubepods-besteffort-pod8bf67c05_04fb_414a_aa6e_8cdc3831cb94.slice" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" Oct 05 07:16:22 crc kubenswrapper[4935]: I1005 07:16:22.655291 4935 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podabe76393-92f2-4164-a549-5f75ced24fb3"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podabe76393-92f2-4164-a549-5f75ced24fb3] : Timed out while waiting for systemd to remove kubepods-besteffort-podabe76393_92f2_4164_a549_5f75ced24fb3.slice" Oct 05 07:16:22 crc kubenswrapper[4935]: E1005 07:16:22.655400 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podabe76393-92f2-4164-a549-5f75ced24fb3] : unable to destroy cgroup paths for cgroup [kubepods besteffort podabe76393-92f2-4164-a549-5f75ced24fb3] : Timed out while waiting for systemd to remove kubepods-besteffort-podabe76393_92f2_4164_a549_5f75ced24fb3.slice" pod="openstack/openstackclient" podUID="abe76393-92f2-4164-a549-5f75ced24fb3" Oct 05 07:16:22 crc kubenswrapper[4935]: I1005 07:16:22.786679 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" path="/var/lib/kubelet/pods/10a5c868-ffd7-44de-8f47-feb4c0ce9121/volumes" Oct 05 07:16:23 crc kubenswrapper[4935]: I1005 07:16:23.256755 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:16:23 crc kubenswrapper[4935]: I1005 07:16:23.256818 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-47z4r" Oct 05 07:16:23 crc kubenswrapper[4935]: I1005 07:16:23.295257 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-47z4r"] Oct 05 07:16:23 crc kubenswrapper[4935]: I1005 07:16:23.303191 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-47z4r"] Oct 05 07:16:23 crc kubenswrapper[4935]: I1005 07:16:23.595494 4935 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podf950314f-27aa-4a44-b13a-4b4f3a7495ab"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podf950314f-27aa-4a44-b13a-4b4f3a7495ab] : Timed out while waiting for systemd to remove kubepods-besteffort-podf950314f_27aa_4a44_b13a_4b4f3a7495ab.slice" Oct 05 07:16:23 crc kubenswrapper[4935]: E1005 07:16:23.596140 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podf950314f-27aa-4a44-b13a-4b4f3a7495ab] : unable to destroy cgroup paths for cgroup [kubepods besteffort podf950314f-27aa-4a44-b13a-4b4f3a7495ab] : Timed out while waiting for systemd to remove kubepods-besteffort-podf950314f_27aa_4a44_b13a_4b4f3a7495ab.slice" pod="openstack/nova-cell1-novncproxy-0" podUID="f950314f-27aa-4a44-b13a-4b4f3a7495ab" Oct 05 07:16:23 crc kubenswrapper[4935]: I1005 07:16:23.599709 4935 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod0095740d-765a-4cba-b0c8-d36592fad800"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod0095740d-765a-4cba-b0c8-d36592fad800] : Timed out while waiting for systemd to remove kubepods-besteffort-pod0095740d_765a_4cba_b0c8_d36592fad800.slice" Oct 05 07:16:23 crc kubenswrapper[4935]: E1005 07:16:23.599769 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod0095740d-765a-4cba-b0c8-d36592fad800] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod0095740d-765a-4cba-b0c8-d36592fad800] : Timed out while waiting for systemd to remove kubepods-besteffort-pod0095740d_765a_4cba_b0c8_d36592fad800.slice" pod="openstack/placementdd1f-account-delete-cgh8g" podUID="0095740d-765a-4cba-b0c8-d36592fad800" Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.264326 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.264339 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementdd1f-account-delete-cgh8g" Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.288340 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementdd1f-account-delete-cgh8g"] Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.299950 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementdd1f-account-delete-cgh8g"] Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.306859 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.311834 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.789000 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0095740d-765a-4cba-b0c8-d36592fad800" path="/var/lib/kubelet/pods/0095740d-765a-4cba-b0c8-d36592fad800/volumes" Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.790920 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf67c05-04fb-414a-aa6e-8cdc3831cb94" path="/var/lib/kubelet/pods/8bf67c05-04fb-414a-aa6e-8cdc3831cb94/volumes" Oct 05 07:16:24 crc kubenswrapper[4935]: I1005 07:16:24.794553 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f950314f-27aa-4a44-b13a-4b4f3a7495ab" path="/var/lib/kubelet/pods/f950314f-27aa-4a44-b13a-4b4f3a7495ab/volumes" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.114150 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell09d1d-account-delete-24dn6" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.199004 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanb178-account-delete-rnqtn" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.205188 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancee563-account-delete-dqlrz" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.218262 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder9bb0-account-delete-8m44f" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.231475 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi70bb-account-delete-5wwbr" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.250883 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbsh2\" (UniqueName: \"kubernetes.io/projected/7143940e-eedf-48a0-b0dc-4ec91df08ad0-kube-api-access-zbsh2\") pod \"7143940e-eedf-48a0-b0dc-4ec91df08ad0\" (UID: \"7143940e-eedf-48a0-b0dc-4ec91df08ad0\") " Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.259062 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7143940e-eedf-48a0-b0dc-4ec91df08ad0-kube-api-access-zbsh2" (OuterVolumeSpecName: "kube-api-access-zbsh2") pod "7143940e-eedf-48a0-b0dc-4ec91df08ad0" (UID: "7143940e-eedf-48a0-b0dc-4ec91df08ad0"). InnerVolumeSpecName "kube-api-access-zbsh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.278762 4935 generic.go:334] "Generic (PLEG): container finished" podID="819c9cf6-f5e5-46af-9fee-c1c4cc88c29a" containerID="b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3" exitCode=137 Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.278841 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi70bb-account-delete-5wwbr" event={"ID":"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a","Type":"ContainerDied","Data":"b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.278853 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi70bb-account-delete-5wwbr" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.278879 4935 scope.go:117] "RemoveContainer" containerID="b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.278868 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi70bb-account-delete-5wwbr" event={"ID":"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a","Type":"ContainerDied","Data":"ea0f16f5c07f8977895c2b0ae19b218ccee8c33ccdd453ed2659ca7da42027fa"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.281323 4935 generic.go:334] "Generic (PLEG): container finished" podID="7143940e-eedf-48a0-b0dc-4ec91df08ad0" containerID="027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9" exitCode=137 Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.281400 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell09d1d-account-delete-24dn6" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.281443 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell09d1d-account-delete-24dn6" event={"ID":"7143940e-eedf-48a0-b0dc-4ec91df08ad0","Type":"ContainerDied","Data":"027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.281515 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell09d1d-account-delete-24dn6" event={"ID":"7143940e-eedf-48a0-b0dc-4ec91df08ad0","Type":"ContainerDied","Data":"292cd469a0e080e6ac3d0a3525828e356bb30257f2ad94878d5ea283b0b91232"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.284128 4935 generic.go:334] "Generic (PLEG): container finished" podID="6bd6e7de-59a5-479d-b2b8-7f060883f6c8" containerID="7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7" exitCode=137 Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.284174 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancee563-account-delete-dqlrz" event={"ID":"6bd6e7de-59a5-479d-b2b8-7f060883f6c8","Type":"ContainerDied","Data":"7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.284193 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancee563-account-delete-dqlrz" event={"ID":"6bd6e7de-59a5-479d-b2b8-7f060883f6c8","Type":"ContainerDied","Data":"6419b79b1d539c5af236dd70c7f6b0764bf7a78b49dfc9d1ca0e2802cc4bae5e"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.284237 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancee563-account-delete-dqlrz" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.288055 4935 generic.go:334] "Generic (PLEG): container finished" podID="b732474b-e717-4d8d-91b7-e00409800f85" containerID="cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd" exitCode=137 Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.288148 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanb178-account-delete-rnqtn" event={"ID":"b732474b-e717-4d8d-91b7-e00409800f85","Type":"ContainerDied","Data":"cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.288182 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanb178-account-delete-rnqtn" event={"ID":"b732474b-e717-4d8d-91b7-e00409800f85","Type":"ContainerDied","Data":"ebe205a0654a7d61ca4c49ac56972ed1cdc7d59980a0fcfd9437cd3e40829de1"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.288234 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanb178-account-delete-rnqtn" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.292634 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder9bb0-account-delete-8m44f" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.292554 4935 generic.go:334] "Generic (PLEG): container finished" podID="420a72a6-506a-42e6-b77a-ee5ca8527768" containerID="f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87" exitCode=137 Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.292688 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder9bb0-account-delete-8m44f" event={"ID":"420a72a6-506a-42e6-b77a-ee5ca8527768","Type":"ContainerDied","Data":"f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.292752 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder9bb0-account-delete-8m44f" event={"ID":"420a72a6-506a-42e6-b77a-ee5ca8527768","Type":"ContainerDied","Data":"5f03f2be56a1ab2560e59e271a14c16c08d5f954a29cb8888f9f85299e881a6e"} Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.305490 4935 scope.go:117] "RemoveContainer" containerID="b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3" Oct 05 07:16:25 crc kubenswrapper[4935]: E1005 07:16:25.306663 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3\": container with ID starting with b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3 not found: ID does not exist" containerID="b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.306714 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3"} err="failed to get container status \"b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3\": rpc error: code = NotFound desc = could not find container \"b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3\": container with ID starting with b8ed537a7ae478290c5bc06ae51f0852bdb7b7fbc7b679479dd20e389ca984e3 not found: ID does not exist" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.306743 4935 scope.go:117] "RemoveContainer" containerID="027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.323208 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell09d1d-account-delete-24dn6"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.330376 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell09d1d-account-delete-24dn6"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.334974 4935 scope.go:117] "RemoveContainer" containerID="027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9" Oct 05 07:16:25 crc kubenswrapper[4935]: E1005 07:16:25.335356 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9\": container with ID starting with 027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9 not found: ID does not exist" containerID="027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.335390 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9"} err="failed to get container status \"027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9\": rpc error: code = NotFound desc = could not find container \"027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9\": container with ID starting with 027a4da8063ed00a9b6dfcd8ea9c01bc44e3bbeb90ecf4233d2114db5591d3f9 not found: ID does not exist" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.335416 4935 scope.go:117] "RemoveContainer" containerID="7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.350300 4935 scope.go:117] "RemoveContainer" containerID="7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7" Oct 05 07:16:25 crc kubenswrapper[4935]: E1005 07:16:25.350610 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7\": container with ID starting with 7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7 not found: ID does not exist" containerID="7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.350643 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7"} err="failed to get container status \"7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7\": rpc error: code = NotFound desc = could not find container \"7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7\": container with ID starting with 7120d911e20f83dc31b1f130e07d1db1e6e99dc89c3e1f0e2c72daef26672dd7 not found: ID does not exist" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.350659 4935 scope.go:117] "RemoveContainer" containerID="cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.353199 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6rwx\" (UniqueName: \"kubernetes.io/projected/6bd6e7de-59a5-479d-b2b8-7f060883f6c8-kube-api-access-j6rwx\") pod \"6bd6e7de-59a5-479d-b2b8-7f060883f6c8\" (UID: \"6bd6e7de-59a5-479d-b2b8-7f060883f6c8\") " Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.353291 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l225h\" (UniqueName: \"kubernetes.io/projected/420a72a6-506a-42e6-b77a-ee5ca8527768-kube-api-access-l225h\") pod \"420a72a6-506a-42e6-b77a-ee5ca8527768\" (UID: \"420a72a6-506a-42e6-b77a-ee5ca8527768\") " Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.353323 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xxfr\" (UniqueName: \"kubernetes.io/projected/819c9cf6-f5e5-46af-9fee-c1c4cc88c29a-kube-api-access-7xxfr\") pod \"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a\" (UID: \"819c9cf6-f5e5-46af-9fee-c1c4cc88c29a\") " Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.353377 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbmhr\" (UniqueName: \"kubernetes.io/projected/b732474b-e717-4d8d-91b7-e00409800f85-kube-api-access-dbmhr\") pod \"b732474b-e717-4d8d-91b7-e00409800f85\" (UID: \"b732474b-e717-4d8d-91b7-e00409800f85\") " Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.353842 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbsh2\" (UniqueName: \"kubernetes.io/projected/7143940e-eedf-48a0-b0dc-4ec91df08ad0-kube-api-access-zbsh2\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.357088 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/819c9cf6-f5e5-46af-9fee-c1c4cc88c29a-kube-api-access-7xxfr" (OuterVolumeSpecName: "kube-api-access-7xxfr") pod "819c9cf6-f5e5-46af-9fee-c1c4cc88c29a" (UID: "819c9cf6-f5e5-46af-9fee-c1c4cc88c29a"). InnerVolumeSpecName "kube-api-access-7xxfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.357204 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b732474b-e717-4d8d-91b7-e00409800f85-kube-api-access-dbmhr" (OuterVolumeSpecName: "kube-api-access-dbmhr") pod "b732474b-e717-4d8d-91b7-e00409800f85" (UID: "b732474b-e717-4d8d-91b7-e00409800f85"). InnerVolumeSpecName "kube-api-access-dbmhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.357808 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bd6e7de-59a5-479d-b2b8-7f060883f6c8-kube-api-access-j6rwx" (OuterVolumeSpecName: "kube-api-access-j6rwx") pod "6bd6e7de-59a5-479d-b2b8-7f060883f6c8" (UID: "6bd6e7de-59a5-479d-b2b8-7f060883f6c8"). InnerVolumeSpecName "kube-api-access-j6rwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.359030 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/420a72a6-506a-42e6-b77a-ee5ca8527768-kube-api-access-l225h" (OuterVolumeSpecName: "kube-api-access-l225h") pod "420a72a6-506a-42e6-b77a-ee5ca8527768" (UID: "420a72a6-506a-42e6-b77a-ee5ca8527768"). InnerVolumeSpecName "kube-api-access-l225h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.365721 4935 scope.go:117] "RemoveContainer" containerID="cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd" Oct 05 07:16:25 crc kubenswrapper[4935]: E1005 07:16:25.366133 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd\": container with ID starting with cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd not found: ID does not exist" containerID="cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.366164 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd"} err="failed to get container status \"cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd\": rpc error: code = NotFound desc = could not find container \"cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd\": container with ID starting with cf612dd47676f86726cf4e03c8f96dee7232a8db8f0efbdfe2dad62af406c3fd not found: ID does not exist" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.366182 4935 scope.go:117] "RemoveContainer" containerID="f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.381748 4935 scope.go:117] "RemoveContainer" containerID="f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87" Oct 05 07:16:25 crc kubenswrapper[4935]: E1005 07:16:25.382258 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87\": container with ID starting with f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87 not found: ID does not exist" containerID="f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.382291 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87"} err="failed to get container status \"f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87\": rpc error: code = NotFound desc = could not find container \"f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87\": container with ID starting with f2edda1f09702521ad09f097394fdf7987c72c1e89c1feaf0e9d317aa0f1ed87 not found: ID does not exist" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.454889 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6rwx\" (UniqueName: \"kubernetes.io/projected/6bd6e7de-59a5-479d-b2b8-7f060883f6c8-kube-api-access-j6rwx\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.454950 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l225h\" (UniqueName: \"kubernetes.io/projected/420a72a6-506a-42e6-b77a-ee5ca8527768-kube-api-access-l225h\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.454960 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xxfr\" (UniqueName: \"kubernetes.io/projected/819c9cf6-f5e5-46af-9fee-c1c4cc88c29a-kube-api-access-7xxfr\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.454968 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbmhr\" (UniqueName: \"kubernetes.io/projected/b732474b-e717-4d8d-91b7-e00409800f85-kube-api-access-dbmhr\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.620986 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi70bb-account-delete-5wwbr"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.626349 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi70bb-account-delete-5wwbr"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.631630 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancee563-account-delete-dqlrz"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.637245 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancee563-account-delete-dqlrz"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.641546 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanb178-account-delete-rnqtn"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.646938 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicanb178-account-delete-rnqtn"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.651319 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder9bb0-account-delete-8m44f"] Oct 05 07:16:25 crc kubenswrapper[4935]: I1005 07:16:25.655099 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder9bb0-account-delete-8m44f"] Oct 05 07:16:26 crc kubenswrapper[4935]: I1005 07:16:26.795284 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="420a72a6-506a-42e6-b77a-ee5ca8527768" path="/var/lib/kubelet/pods/420a72a6-506a-42e6-b77a-ee5ca8527768/volumes" Oct 05 07:16:26 crc kubenswrapper[4935]: I1005 07:16:26.796511 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bd6e7de-59a5-479d-b2b8-7f060883f6c8" path="/var/lib/kubelet/pods/6bd6e7de-59a5-479d-b2b8-7f060883f6c8/volumes" Oct 05 07:16:26 crc kubenswrapper[4935]: I1005 07:16:26.797604 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7143940e-eedf-48a0-b0dc-4ec91df08ad0" path="/var/lib/kubelet/pods/7143940e-eedf-48a0-b0dc-4ec91df08ad0/volumes" Oct 05 07:16:26 crc kubenswrapper[4935]: I1005 07:16:26.798694 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="819c9cf6-f5e5-46af-9fee-c1c4cc88c29a" path="/var/lib/kubelet/pods/819c9cf6-f5e5-46af-9fee-c1c4cc88c29a/volumes" Oct 05 07:16:26 crc kubenswrapper[4935]: I1005 07:16:26.800632 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b732474b-e717-4d8d-91b7-e00409800f85" path="/var/lib/kubelet/pods/b732474b-e717-4d8d-91b7-e00409800f85/volumes" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.000206 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wtmnn"] Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.000991 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819c9cf6-f5e5-46af-9fee-c1c4cc88c29a" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001007 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="819c9cf6-f5e5-46af-9fee-c1c4cc88c29a" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001017 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001023 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001039 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001045 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001055 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-reaper" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001062 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-reaper" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001076 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-updater" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001082 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-updater" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001096 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001104 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-server" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001113 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b732474b-e717-4d8d-91b7-e00409800f85" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001119 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b732474b-e717-4d8d-91b7-e00409800f85" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001133 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001139 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001149 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27a28306-70be-4556-9659-64999f775195" containerName="neutron-httpd" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001157 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="27a28306-70be-4556-9659-64999f775195" containerName="neutron-httpd" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001169 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server-init" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001175 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server-init" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001184 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-expirer" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001190 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-expirer" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001198 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerName="registry-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001204 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerName="registry-server" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001217 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001222 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-server" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001231 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="swift-recon-cron" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001237 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="swift-recon-cron" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001248 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27a28306-70be-4556-9659-64999f775195" containerName="neutron-api" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001253 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="27a28306-70be-4556-9659-64999f775195" containerName="neutron-api" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001263 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerName="extract-utilities" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001270 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerName="extract-utilities" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001277 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001282 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001292 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420a72a6-506a-42e6-b77a-ee5ca8527768" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001298 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="420a72a6-506a-42e6-b77a-ee5ca8527768" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001305 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="rsync" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001311 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="rsync" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001318 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001327 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001338 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001346 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-server" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001359 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001367 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001380 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001387 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001399 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001406 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001417 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerName="extract-content" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001424 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerName="extract-content" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001431 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd6e7de-59a5-479d-b2b8-7f060883f6c8" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001437 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd6e7de-59a5-479d-b2b8-7f060883f6c8" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001448 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-updater" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001453 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-updater" Oct 05 07:16:38 crc kubenswrapper[4935]: E1005 07:16:38.001462 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7143940e-eedf-48a0-b0dc-4ec91df08ad0" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001467 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7143940e-eedf-48a0-b0dc-4ec91df08ad0" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001588 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovsdb-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001597 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac8194b-9784-4eed-8c35-fb65a584b525" containerName="ovs-vswitchd" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001603 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7143940e-eedf-48a0-b0dc-4ec91df08ad0" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001614 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-reaper" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001625 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001634 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001645 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-expirer" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001653 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="rsync" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001662 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001673 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001682 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001691 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="27a28306-70be-4556-9659-64999f775195" containerName="neutron-httpd" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001701 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001709 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="swift-recon-cron" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001717 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b732474b-e717-4d8d-91b7-e00409800f85" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001730 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="27a28306-70be-4556-9659-64999f775195" containerName="neutron-api" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001737 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="account-auditor" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001745 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="420a72a6-506a-42e6-b77a-ee5ca8527768" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001754 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-updater" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001761 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001768 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd6e7de-59a5-479d-b2b8-7f060883f6c8" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001775 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="819c9cf6-f5e5-46af-9fee-c1c4cc88c29a" containerName="mariadb-account-delete" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001784 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="container-replicator" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001793 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b9e1fa-5995-4685-9d9c-e38469c4e7a1" containerName="registry-server" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.001802 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a5c868-ffd7-44de-8f47-feb4c0ce9121" containerName="object-updater" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.002857 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.069072 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wtmnn"] Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.141448 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-catalog-content\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.141512 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6lrf\" (UniqueName: \"kubernetes.io/projected/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-kube-api-access-z6lrf\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.141726 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-utilities\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.243230 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-utilities\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.243292 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-catalog-content\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.243321 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6lrf\" (UniqueName: \"kubernetes.io/projected/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-kube-api-access-z6lrf\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.244060 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-utilities\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.244114 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-catalog-content\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.280438 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6lrf\" (UniqueName: \"kubernetes.io/projected/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-kube-api-access-z6lrf\") pod \"certified-operators-wtmnn\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.342178 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:38 crc kubenswrapper[4935]: I1005 07:16:38.840234 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wtmnn"] Oct 05 07:16:39 crc kubenswrapper[4935]: I1005 07:16:39.465822 4935 generic.go:334] "Generic (PLEG): container finished" podID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerID="e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8" exitCode=0 Oct 05 07:16:39 crc kubenswrapper[4935]: I1005 07:16:39.465926 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtmnn" event={"ID":"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4","Type":"ContainerDied","Data":"e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8"} Oct 05 07:16:39 crc kubenswrapper[4935]: I1005 07:16:39.465976 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtmnn" event={"ID":"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4","Type":"ContainerStarted","Data":"349342efc5b0fbaf1386ca6b7d8c6c0540bb1fed13f6aff0b3a22169c578ae80"} Oct 05 07:16:40 crc kubenswrapper[4935]: I1005 07:16:40.477479 4935 generic.go:334] "Generic (PLEG): container finished" podID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerID="bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281" exitCode=0 Oct 05 07:16:40 crc kubenswrapper[4935]: I1005 07:16:40.477654 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtmnn" event={"ID":"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4","Type":"ContainerDied","Data":"bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281"} Oct 05 07:16:41 crc kubenswrapper[4935]: I1005 07:16:41.489235 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtmnn" event={"ID":"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4","Type":"ContainerStarted","Data":"c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b"} Oct 05 07:16:41 crc kubenswrapper[4935]: I1005 07:16:41.520993 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wtmnn" podStartSLOduration=3.115458212 podStartE2EDuration="4.520973149s" podCreationTimestamp="2025-10-05 07:16:37 +0000 UTC" firstStartedPulling="2025-10-05 07:16:39.468864725 +0000 UTC m=+1433.351491215" lastFinishedPulling="2025-10-05 07:16:40.874379692 +0000 UTC m=+1434.757006152" observedRunningTime="2025-10-05 07:16:41.51531905 +0000 UTC m=+1435.397945510" watchObservedRunningTime="2025-10-05 07:16:41.520973149 +0000 UTC m=+1435.403599609" Oct 05 07:16:44 crc kubenswrapper[4935]: I1005 07:16:44.289820 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:16:44 crc kubenswrapper[4935]: I1005 07:16:44.290227 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:16:48 crc kubenswrapper[4935]: I1005 07:16:48.343496 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:48 crc kubenswrapper[4935]: I1005 07:16:48.343866 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:48 crc kubenswrapper[4935]: I1005 07:16:48.386377 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:48 crc kubenswrapper[4935]: I1005 07:16:48.599272 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:48 crc kubenswrapper[4935]: I1005 07:16:48.648302 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wtmnn"] Oct 05 07:16:50 crc kubenswrapper[4935]: I1005 07:16:50.566261 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wtmnn" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerName="registry-server" containerID="cri-o://c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b" gracePeriod=2 Oct 05 07:16:50 crc kubenswrapper[4935]: E1005 07:16:50.888471 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e6764cc_271a_42e1_a3cb_3c9fa0d3f2d4.slice/crio-conmon-c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b.scope\": RecentStats: unable to find data in memory cache]" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.529469 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.576333 4935 generic.go:334] "Generic (PLEG): container finished" podID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerID="c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b" exitCode=0 Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.576376 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtmnn" event={"ID":"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4","Type":"ContainerDied","Data":"c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b"} Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.576404 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wtmnn" event={"ID":"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4","Type":"ContainerDied","Data":"349342efc5b0fbaf1386ca6b7d8c6c0540bb1fed13f6aff0b3a22169c578ae80"} Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.576424 4935 scope.go:117] "RemoveContainer" containerID="c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.576446 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wtmnn" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.619760 4935 scope.go:117] "RemoveContainer" containerID="bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.653565 4935 scope.go:117] "RemoveContainer" containerID="e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.654525 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-catalog-content\") pod \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.654637 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-utilities\") pod \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.654726 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6lrf\" (UniqueName: \"kubernetes.io/projected/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-kube-api-access-z6lrf\") pod \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\" (UID: \"0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4\") " Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.657850 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-utilities" (OuterVolumeSpecName: "utilities") pod "0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" (UID: "0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.663744 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-kube-api-access-z6lrf" (OuterVolumeSpecName: "kube-api-access-z6lrf") pod "0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" (UID: "0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4"). InnerVolumeSpecName "kube-api-access-z6lrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.705851 4935 scope.go:117] "RemoveContainer" containerID="c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b" Oct 05 07:16:51 crc kubenswrapper[4935]: E1005 07:16:51.709405 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b\": container with ID starting with c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b not found: ID does not exist" containerID="c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.709452 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b"} err="failed to get container status \"c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b\": rpc error: code = NotFound desc = could not find container \"c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b\": container with ID starting with c70010d324745b40c24c314a33f353ab8589781166f2f3613bde4acd9808b51b not found: ID does not exist" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.709505 4935 scope.go:117] "RemoveContainer" containerID="bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281" Oct 05 07:16:51 crc kubenswrapper[4935]: E1005 07:16:51.710115 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281\": container with ID starting with bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281 not found: ID does not exist" containerID="bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.710158 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281"} err="failed to get container status \"bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281\": rpc error: code = NotFound desc = could not find container \"bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281\": container with ID starting with bca35e5ff79fcdff8feeb9b8eb4f6a1d3d0e513e8a4de61fcbb38d14cfbb4281 not found: ID does not exist" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.710183 4935 scope.go:117] "RemoveContainer" containerID="e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8" Oct 05 07:16:51 crc kubenswrapper[4935]: E1005 07:16:51.710458 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8\": container with ID starting with e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8 not found: ID does not exist" containerID="e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.710730 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8"} err="failed to get container status \"e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8\": rpc error: code = NotFound desc = could not find container \"e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8\": container with ID starting with e76d275275ac8919d908f88bba134301775dce5c18587015f4f043df7f6695e8 not found: ID does not exist" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.718391 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" (UID: "0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.757443 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.757493 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.757513 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6lrf\" (UniqueName: \"kubernetes.io/projected/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4-kube-api-access-z6lrf\") on node \"crc\" DevicePath \"\"" Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.913048 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wtmnn"] Oct 05 07:16:51 crc kubenswrapper[4935]: I1005 07:16:51.921321 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wtmnn"] Oct 05 07:16:52 crc kubenswrapper[4935]: I1005 07:16:52.796185 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" path="/var/lib/kubelet/pods/0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4/volumes" Oct 05 07:16:54 crc kubenswrapper[4935]: I1005 07:16:54.941801 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qgsbl"] Oct 05 07:16:54 crc kubenswrapper[4935]: E1005 07:16:54.942884 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerName="registry-server" Oct 05 07:16:54 crc kubenswrapper[4935]: I1005 07:16:54.942953 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerName="registry-server" Oct 05 07:16:54 crc kubenswrapper[4935]: E1005 07:16:54.943027 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerName="extract-content" Oct 05 07:16:54 crc kubenswrapper[4935]: I1005 07:16:54.943046 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerName="extract-content" Oct 05 07:16:54 crc kubenswrapper[4935]: E1005 07:16:54.943094 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerName="extract-utilities" Oct 05 07:16:54 crc kubenswrapper[4935]: I1005 07:16:54.943112 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerName="extract-utilities" Oct 05 07:16:54 crc kubenswrapper[4935]: I1005 07:16:54.943425 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6764cc-271a-42e1-a3cb-3c9fa0d3f2d4" containerName="registry-server" Oct 05 07:16:54 crc kubenswrapper[4935]: I1005 07:16:54.946056 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:54 crc kubenswrapper[4935]: I1005 07:16:54.958838 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgsbl"] Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.109769 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lht9s\" (UniqueName: \"kubernetes.io/projected/24e495e7-4edc-4d86-9a82-13c33e0b0465-kube-api-access-lht9s\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.110024 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-utilities\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.110129 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-catalog-content\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.211452 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-utilities\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.211520 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-catalog-content\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.211584 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lht9s\" (UniqueName: \"kubernetes.io/projected/24e495e7-4edc-4d86-9a82-13c33e0b0465-kube-api-access-lht9s\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.212027 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-utilities\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.212084 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-catalog-content\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.235661 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lht9s\" (UniqueName: \"kubernetes.io/projected/24e495e7-4edc-4d86-9a82-13c33e0b0465-kube-api-access-lht9s\") pod \"community-operators-qgsbl\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.268718 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:16:55 crc kubenswrapper[4935]: I1005 07:16:55.750764 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgsbl"] Oct 05 07:16:56 crc kubenswrapper[4935]: I1005 07:16:56.640839 4935 generic.go:334] "Generic (PLEG): container finished" podID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerID="a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b" exitCode=0 Oct 05 07:16:56 crc kubenswrapper[4935]: I1005 07:16:56.640992 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgsbl" event={"ID":"24e495e7-4edc-4d86-9a82-13c33e0b0465","Type":"ContainerDied","Data":"a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b"} Oct 05 07:16:56 crc kubenswrapper[4935]: I1005 07:16:56.641112 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgsbl" event={"ID":"24e495e7-4edc-4d86-9a82-13c33e0b0465","Type":"ContainerStarted","Data":"29bc904fcea0d81e5411cf8f54d1e17477286679ae00e32ff4397caa705986b9"} Oct 05 07:16:57 crc kubenswrapper[4935]: I1005 07:16:57.655267 4935 generic.go:334] "Generic (PLEG): container finished" podID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerID="9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3" exitCode=0 Oct 05 07:16:57 crc kubenswrapper[4935]: I1005 07:16:57.655365 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgsbl" event={"ID":"24e495e7-4edc-4d86-9a82-13c33e0b0465","Type":"ContainerDied","Data":"9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3"} Oct 05 07:16:58 crc kubenswrapper[4935]: I1005 07:16:58.672385 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgsbl" event={"ID":"24e495e7-4edc-4d86-9a82-13c33e0b0465","Type":"ContainerStarted","Data":"0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd"} Oct 05 07:16:58 crc kubenswrapper[4935]: I1005 07:16:58.698781 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qgsbl" podStartSLOduration=3.291677452 podStartE2EDuration="4.698755671s" podCreationTimestamp="2025-10-05 07:16:54 +0000 UTC" firstStartedPulling="2025-10-05 07:16:56.643504163 +0000 UTC m=+1450.526130673" lastFinishedPulling="2025-10-05 07:16:58.050582422 +0000 UTC m=+1451.933208892" observedRunningTime="2025-10-05 07:16:58.69872902 +0000 UTC m=+1452.581355510" watchObservedRunningTime="2025-10-05 07:16:58.698755671 +0000 UTC m=+1452.581382171" Oct 05 07:17:05 crc kubenswrapper[4935]: I1005 07:17:05.269864 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:17:05 crc kubenswrapper[4935]: I1005 07:17:05.270350 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:17:05 crc kubenswrapper[4935]: I1005 07:17:05.347127 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:17:05 crc kubenswrapper[4935]: I1005 07:17:05.810034 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:17:05 crc kubenswrapper[4935]: I1005 07:17:05.863297 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgsbl"] Oct 05 07:17:07 crc kubenswrapper[4935]: I1005 07:17:07.761836 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qgsbl" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerName="registry-server" containerID="cri-o://0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd" gracePeriod=2 Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.193733 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.319813 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-catalog-content\") pod \"24e495e7-4edc-4d86-9a82-13c33e0b0465\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.320019 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lht9s\" (UniqueName: \"kubernetes.io/projected/24e495e7-4edc-4d86-9a82-13c33e0b0465-kube-api-access-lht9s\") pod \"24e495e7-4edc-4d86-9a82-13c33e0b0465\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.320111 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-utilities\") pod \"24e495e7-4edc-4d86-9a82-13c33e0b0465\" (UID: \"24e495e7-4edc-4d86-9a82-13c33e0b0465\") " Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.321714 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-utilities" (OuterVolumeSpecName: "utilities") pod "24e495e7-4edc-4d86-9a82-13c33e0b0465" (UID: "24e495e7-4edc-4d86-9a82-13c33e0b0465"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.325421 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24e495e7-4edc-4d86-9a82-13c33e0b0465-kube-api-access-lht9s" (OuterVolumeSpecName: "kube-api-access-lht9s") pod "24e495e7-4edc-4d86-9a82-13c33e0b0465" (UID: "24e495e7-4edc-4d86-9a82-13c33e0b0465"). InnerVolumeSpecName "kube-api-access-lht9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.393141 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24e495e7-4edc-4d86-9a82-13c33e0b0465" (UID: "24e495e7-4edc-4d86-9a82-13c33e0b0465"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.423024 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lht9s\" (UniqueName: \"kubernetes.io/projected/24e495e7-4edc-4d86-9a82-13c33e0b0465-kube-api-access-lht9s\") on node \"crc\" DevicePath \"\"" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.423060 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.423072 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24e495e7-4edc-4d86-9a82-13c33e0b0465-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.775972 4935 generic.go:334] "Generic (PLEG): container finished" podID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerID="0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd" exitCode=0 Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.776079 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgsbl" event={"ID":"24e495e7-4edc-4d86-9a82-13c33e0b0465","Type":"ContainerDied","Data":"0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd"} Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.776166 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgsbl" event={"ID":"24e495e7-4edc-4d86-9a82-13c33e0b0465","Type":"ContainerDied","Data":"29bc904fcea0d81e5411cf8f54d1e17477286679ae00e32ff4397caa705986b9"} Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.776199 4935 scope.go:117] "RemoveContainer" containerID="0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.776231 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgsbl" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.813576 4935 scope.go:117] "RemoveContainer" containerID="9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.840233 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgsbl"] Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.850125 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qgsbl"] Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.862988 4935 scope.go:117] "RemoveContainer" containerID="a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.890402 4935 scope.go:117] "RemoveContainer" containerID="0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd" Oct 05 07:17:08 crc kubenswrapper[4935]: E1005 07:17:08.890738 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd\": container with ID starting with 0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd not found: ID does not exist" containerID="0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.890776 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd"} err="failed to get container status \"0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd\": rpc error: code = NotFound desc = could not find container \"0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd\": container with ID starting with 0970e206a5d181562fd70baff46623b79ece101b0bd0768d7c7faacc804277fd not found: ID does not exist" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.890804 4935 scope.go:117] "RemoveContainer" containerID="9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3" Oct 05 07:17:08 crc kubenswrapper[4935]: E1005 07:17:08.891142 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3\": container with ID starting with 9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3 not found: ID does not exist" containerID="9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.891165 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3"} err="failed to get container status \"9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3\": rpc error: code = NotFound desc = could not find container \"9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3\": container with ID starting with 9318e1fa7ffac3176385a02cc7d3c43c41cd981d409a25060eedec8dd8820ff3 not found: ID does not exist" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.891179 4935 scope.go:117] "RemoveContainer" containerID="a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b" Oct 05 07:17:08 crc kubenswrapper[4935]: E1005 07:17:08.891385 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b\": container with ID starting with a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b not found: ID does not exist" containerID="a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b" Oct 05 07:17:08 crc kubenswrapper[4935]: I1005 07:17:08.891422 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b"} err="failed to get container status \"a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b\": rpc error: code = NotFound desc = could not find container \"a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b\": container with ID starting with a0d4603ad92b64833aaa3edb1ef95dd0271165ee7dac39650e311c6135341e8b not found: ID does not exist" Oct 05 07:17:10 crc kubenswrapper[4935]: I1005 07:17:10.789973 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" path="/var/lib/kubelet/pods/24e495e7-4edc-4d86-9a82-13c33e0b0465/volumes" Oct 05 07:17:14 crc kubenswrapper[4935]: I1005 07:17:14.289752 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:17:14 crc kubenswrapper[4935]: I1005 07:17:14.290265 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:17:44 crc kubenswrapper[4935]: I1005 07:17:44.289678 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:17:44 crc kubenswrapper[4935]: I1005 07:17:44.290566 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:17:44 crc kubenswrapper[4935]: I1005 07:17:44.290650 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:17:44 crc kubenswrapper[4935]: I1005 07:17:44.291768 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:17:44 crc kubenswrapper[4935]: I1005 07:17:44.291877 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" gracePeriod=600 Oct 05 07:17:44 crc kubenswrapper[4935]: E1005 07:17:44.434078 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:17:45 crc kubenswrapper[4935]: I1005 07:17:45.256876 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" exitCode=0 Oct 05 07:17:45 crc kubenswrapper[4935]: I1005 07:17:45.256945 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072"} Oct 05 07:17:45 crc kubenswrapper[4935]: I1005 07:17:45.257015 4935 scope.go:117] "RemoveContainer" containerID="7b6d2f4c73d4ef3b7b40cc3341e20f2c6b3b4fe3d7affc8f4b8588270c63e170" Oct 05 07:17:45 crc kubenswrapper[4935]: I1005 07:17:45.258123 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:17:45 crc kubenswrapper[4935]: E1005 07:17:45.258639 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.060703 4935 scope.go:117] "RemoveContainer" containerID="24c80e7ed845e73bf331fc87b91c537868571e5394f6a48ad9f2db5a21d684ab" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.083645 4935 scope.go:117] "RemoveContainer" containerID="158202d939bff1d98cf1e844dd1d81256f5a1daac2a99088ace4f7eb6d20c664" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.156540 4935 scope.go:117] "RemoveContainer" containerID="acc87771f402f72e5094e197427a1e741957a406635fe722bbeda86770f046c0" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.277284 4935 scope.go:117] "RemoveContainer" containerID="c0b64cde633e9f03356159545948baedda001bcfee42e911ce7189993ae44f3a" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.308444 4935 scope.go:117] "RemoveContainer" containerID="28bbaf182e3a7272bba91cd40617c6d28d901e11a85bd3ebc75a3a8af669c7b2" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.325322 4935 scope.go:117] "RemoveContainer" containerID="adfb600248e13d2f6b1320de3c5a3560bd39190a465558e7111ae839f318c675" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.350409 4935 scope.go:117] "RemoveContainer" containerID="5985a42517f3f6b261d93f0ab627c53c5bc297b9e7699a332b44b1e66d30ab31" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.371514 4935 scope.go:117] "RemoveContainer" containerID="9f89c3a737898aa877917fa322569786f07234b97e0640a587d32d6f00f43018" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.390977 4935 scope.go:117] "RemoveContainer" containerID="d7c7e55c571cb0cfdbd49c39af5fe9080b3f370254ae163b6d61b41c9d8ee5cf" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.406692 4935 scope.go:117] "RemoveContainer" containerID="a8d4be873da7a3591175629c55ecf97d564e67d03648bff86a6511dce9218fce" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.422737 4935 scope.go:117] "RemoveContainer" containerID="b721fb230c38e8b61034175f3484540666a622c814025c89af60628e2f544e6d" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.437432 4935 scope.go:117] "RemoveContainer" containerID="0f88e2c22830dd1c5d7712c2f78dedffa2c6264f0bb2ea8d8f08ba4a069716c1" Oct 05 07:17:52 crc kubenswrapper[4935]: I1005 07:17:52.459852 4935 scope.go:117] "RemoveContainer" containerID="183299868f8eadbbb5a725d9ebfc79c7434ded372ecb00a64b395cdd724042a7" Oct 05 07:17:58 crc kubenswrapper[4935]: I1005 07:17:58.778794 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:17:58 crc kubenswrapper[4935]: E1005 07:17:58.779962 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:18:10 crc kubenswrapper[4935]: I1005 07:18:10.777074 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:18:10 crc kubenswrapper[4935]: E1005 07:18:10.777954 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:18:25 crc kubenswrapper[4935]: I1005 07:18:25.777984 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:18:25 crc kubenswrapper[4935]: E1005 07:18:25.778663 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.290744 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-scvwl"] Oct 05 07:18:30 crc kubenswrapper[4935]: E1005 07:18:30.291735 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerName="registry-server" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.291772 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerName="registry-server" Oct 05 07:18:30 crc kubenswrapper[4935]: E1005 07:18:30.291812 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerName="extract-utilities" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.291830 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerName="extract-utilities" Oct 05 07:18:30 crc kubenswrapper[4935]: E1005 07:18:30.291883 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerName="extract-content" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.291935 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerName="extract-content" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.292218 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="24e495e7-4edc-4d86-9a82-13c33e0b0465" containerName="registry-server" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.294261 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.306461 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-scvwl"] Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.443166 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7tfp\" (UniqueName: \"kubernetes.io/projected/adb043d5-2816-4b9f-82b0-1f89170c0cdc-kube-api-access-g7tfp\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.443213 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-utilities\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.443314 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-catalog-content\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.544577 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7tfp\" (UniqueName: \"kubernetes.io/projected/adb043d5-2816-4b9f-82b0-1f89170c0cdc-kube-api-access-g7tfp\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.544657 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-utilities\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.544714 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-catalog-content\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.545295 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-utilities\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.545406 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-catalog-content\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.571575 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7tfp\" (UniqueName: \"kubernetes.io/projected/adb043d5-2816-4b9f-82b0-1f89170c0cdc-kube-api-access-g7tfp\") pod \"redhat-marketplace-scvwl\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:30 crc kubenswrapper[4935]: I1005 07:18:30.627451 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:31 crc kubenswrapper[4935]: I1005 07:18:31.082820 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-scvwl"] Oct 05 07:18:31 crc kubenswrapper[4935]: I1005 07:18:31.699809 4935 generic.go:334] "Generic (PLEG): container finished" podID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerID="992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481" exitCode=0 Oct 05 07:18:31 crc kubenswrapper[4935]: I1005 07:18:31.699992 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scvwl" event={"ID":"adb043d5-2816-4b9f-82b0-1f89170c0cdc","Type":"ContainerDied","Data":"992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481"} Oct 05 07:18:31 crc kubenswrapper[4935]: I1005 07:18:31.700159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scvwl" event={"ID":"adb043d5-2816-4b9f-82b0-1f89170c0cdc","Type":"ContainerStarted","Data":"5373d31d7b3957e3f863d553db16958371cfda5fa47a3cec4eaf8b3d8d5d6c1c"} Oct 05 07:18:32 crc kubenswrapper[4935]: I1005 07:18:32.709406 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scvwl" event={"ID":"adb043d5-2816-4b9f-82b0-1f89170c0cdc","Type":"ContainerDied","Data":"4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d"} Oct 05 07:18:32 crc kubenswrapper[4935]: I1005 07:18:32.709266 4935 generic.go:334] "Generic (PLEG): container finished" podID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerID="4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d" exitCode=0 Oct 05 07:18:33 crc kubenswrapper[4935]: I1005 07:18:33.722171 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scvwl" event={"ID":"adb043d5-2816-4b9f-82b0-1f89170c0cdc","Type":"ContainerStarted","Data":"917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb"} Oct 05 07:18:33 crc kubenswrapper[4935]: I1005 07:18:33.747429 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-scvwl" podStartSLOduration=2.359020821 podStartE2EDuration="3.747398063s" podCreationTimestamp="2025-10-05 07:18:30 +0000 UTC" firstStartedPulling="2025-10-05 07:18:31.70285535 +0000 UTC m=+1545.585481850" lastFinishedPulling="2025-10-05 07:18:33.091232642 +0000 UTC m=+1546.973859092" observedRunningTime="2025-10-05 07:18:33.742153534 +0000 UTC m=+1547.624780064" watchObservedRunningTime="2025-10-05 07:18:33.747398063 +0000 UTC m=+1547.630024553" Oct 05 07:18:37 crc kubenswrapper[4935]: I1005 07:18:37.777622 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:18:37 crc kubenswrapper[4935]: E1005 07:18:37.778341 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:18:40 crc kubenswrapper[4935]: I1005 07:18:40.628368 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:40 crc kubenswrapper[4935]: I1005 07:18:40.628771 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:40 crc kubenswrapper[4935]: I1005 07:18:40.684576 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:40 crc kubenswrapper[4935]: I1005 07:18:40.880963 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:40 crc kubenswrapper[4935]: I1005 07:18:40.938729 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-scvwl"] Oct 05 07:18:42 crc kubenswrapper[4935]: I1005 07:18:42.809790 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-scvwl" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerName="registry-server" containerID="cri-o://917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb" gracePeriod=2 Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.216491 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.346481 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7tfp\" (UniqueName: \"kubernetes.io/projected/adb043d5-2816-4b9f-82b0-1f89170c0cdc-kube-api-access-g7tfp\") pod \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.346555 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-catalog-content\") pod \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.346611 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-utilities\") pod \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\" (UID: \"adb043d5-2816-4b9f-82b0-1f89170c0cdc\") " Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.347796 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-utilities" (OuterVolumeSpecName: "utilities") pod "adb043d5-2816-4b9f-82b0-1f89170c0cdc" (UID: "adb043d5-2816-4b9f-82b0-1f89170c0cdc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.352370 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adb043d5-2816-4b9f-82b0-1f89170c0cdc-kube-api-access-g7tfp" (OuterVolumeSpecName: "kube-api-access-g7tfp") pod "adb043d5-2816-4b9f-82b0-1f89170c0cdc" (UID: "adb043d5-2816-4b9f-82b0-1f89170c0cdc"). InnerVolumeSpecName "kube-api-access-g7tfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.368835 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adb043d5-2816-4b9f-82b0-1f89170c0cdc" (UID: "adb043d5-2816-4b9f-82b0-1f89170c0cdc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.448736 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7tfp\" (UniqueName: \"kubernetes.io/projected/adb043d5-2816-4b9f-82b0-1f89170c0cdc-kube-api-access-g7tfp\") on node \"crc\" DevicePath \"\"" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.448786 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.448807 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adb043d5-2816-4b9f-82b0-1f89170c0cdc-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.824411 4935 generic.go:334] "Generic (PLEG): container finished" podID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerID="917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb" exitCode=0 Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.824473 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scvwl" event={"ID":"adb043d5-2816-4b9f-82b0-1f89170c0cdc","Type":"ContainerDied","Data":"917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb"} Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.824515 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scvwl" event={"ID":"adb043d5-2816-4b9f-82b0-1f89170c0cdc","Type":"ContainerDied","Data":"5373d31d7b3957e3f863d553db16958371cfda5fa47a3cec4eaf8b3d8d5d6c1c"} Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.824531 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scvwl" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.824565 4935 scope.go:117] "RemoveContainer" containerID="917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.860031 4935 scope.go:117] "RemoveContainer" containerID="4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.877134 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-scvwl"] Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.887094 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-scvwl"] Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.899380 4935 scope.go:117] "RemoveContainer" containerID="992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.930592 4935 scope.go:117] "RemoveContainer" containerID="917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb" Oct 05 07:18:43 crc kubenswrapper[4935]: E1005 07:18:43.931214 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb\": container with ID starting with 917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb not found: ID does not exist" containerID="917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.931286 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb"} err="failed to get container status \"917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb\": rpc error: code = NotFound desc = could not find container \"917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb\": container with ID starting with 917b798fae34083c8058231849e773be18c7428a4145b4d3d3cea2bf7aef2fcb not found: ID does not exist" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.931325 4935 scope.go:117] "RemoveContainer" containerID="4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d" Oct 05 07:18:43 crc kubenswrapper[4935]: E1005 07:18:43.931733 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d\": container with ID starting with 4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d not found: ID does not exist" containerID="4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.931772 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d"} err="failed to get container status \"4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d\": rpc error: code = NotFound desc = could not find container \"4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d\": container with ID starting with 4901be7a0f74a01c20bc3235253c4071c90033a42179d3d7a035951ab025452d not found: ID does not exist" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.931798 4935 scope.go:117] "RemoveContainer" containerID="992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481" Oct 05 07:18:43 crc kubenswrapper[4935]: E1005 07:18:43.932127 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481\": container with ID starting with 992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481 not found: ID does not exist" containerID="992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481" Oct 05 07:18:43 crc kubenswrapper[4935]: I1005 07:18:43.932187 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481"} err="failed to get container status \"992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481\": rpc error: code = NotFound desc = could not find container \"992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481\": container with ID starting with 992a426b621c2ecc08db50867a5afdf8a20684fdf30b42badc88281a00256481 not found: ID does not exist" Oct 05 07:18:44 crc kubenswrapper[4935]: I1005 07:18:44.794232 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" path="/var/lib/kubelet/pods/adb043d5-2816-4b9f-82b0-1f89170c0cdc/volumes" Oct 05 07:18:48 crc kubenswrapper[4935]: I1005 07:18:48.777401 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:18:48 crc kubenswrapper[4935]: E1005 07:18:48.778239 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.711935 4935 scope.go:117] "RemoveContainer" containerID="dae10e0c5cc975b8622037342618fdb57f961bf238a26977922a4a5038cc681f" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.746230 4935 scope.go:117] "RemoveContainer" containerID="221fe2b717870808d7d0e598673cb75e96ed270806b646d7a5d1848b41337223" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.775335 4935 scope.go:117] "RemoveContainer" containerID="f72801a860065d79604fdd54e0db50ff9192f050ef2dd2fefaa717d191b2c075" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.828627 4935 scope.go:117] "RemoveContainer" containerID="7415798dcf5ca1b35b5ba66c1a74be46595f88e290eaa7b69d58158f804e237f" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.862914 4935 scope.go:117] "RemoveContainer" containerID="068f0258b4f42367cfe09495e247e8ae26cd3ea724102859b0fd3528a91385e1" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.886807 4935 scope.go:117] "RemoveContainer" containerID="7cf0b54a3199310d1b75ab3b5890bce9c54028af4fb2c5cde26e65b39f123076" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.907819 4935 scope.go:117] "RemoveContainer" containerID="4ac38cab9550b15f03bb7665210b5060057bc358f157b890ebd6d4e945ea3261" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.944124 4935 scope.go:117] "RemoveContainer" containerID="e61a98d7b38be24992c47d287153ee58d5df1cadd58c8589fb3c0f806844634c" Oct 05 07:18:52 crc kubenswrapper[4935]: I1005 07:18:52.984490 4935 scope.go:117] "RemoveContainer" containerID="5921e4650b95f588079eb80fa686b2459174b8c9e94cdf90d7ec54954b6b541d" Oct 05 07:18:53 crc kubenswrapper[4935]: I1005 07:18:53.013042 4935 scope.go:117] "RemoveContainer" containerID="4b10ad46cc6f92c33279323585a2011416f051915eeaaf3790fc0bd4cbf3080f" Oct 05 07:18:53 crc kubenswrapper[4935]: I1005 07:18:53.038358 4935 scope.go:117] "RemoveContainer" containerID="8b173ce4d56c1415ae5822233d48c1bb3df1e36d7820b9e23201e8af34a167ad" Oct 05 07:18:59 crc kubenswrapper[4935]: I1005 07:18:59.778403 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:18:59 crc kubenswrapper[4935]: E1005 07:18:59.779610 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:19:11 crc kubenswrapper[4935]: I1005 07:19:11.777342 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:19:11 crc kubenswrapper[4935]: E1005 07:19:11.778462 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:19:24 crc kubenswrapper[4935]: I1005 07:19:24.777046 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:19:24 crc kubenswrapper[4935]: E1005 07:19:24.777706 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:19:39 crc kubenswrapper[4935]: I1005 07:19:39.777731 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:19:39 crc kubenswrapper[4935]: E1005 07:19:39.778812 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:19:52 crc kubenswrapper[4935]: I1005 07:19:52.777538 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:19:52 crc kubenswrapper[4935]: E1005 07:19:52.778276 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.275800 4935 scope.go:117] "RemoveContainer" containerID="119888a24ef912962cb413c88d9903d21aad373b3de77a30cfb8ecc482ff8f75" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.317181 4935 scope.go:117] "RemoveContainer" containerID="b4761f2806893419fa049bdae0c7dd84eb9621a5db9d1c6aa5094759cbd94a58" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.353863 4935 scope.go:117] "RemoveContainer" containerID="afd975882d932c2ed56cbd1069d50977a2d2abf98e1305a15671c213fc965de8" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.387513 4935 scope.go:117] "RemoveContainer" containerID="1226b9b0e256c9e620dc56bf185da59022a8cb075c026e4bc864757ee53d82ec" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.421101 4935 scope.go:117] "RemoveContainer" containerID="20a20a4e69bdd787df49c91056089308888f8504fc31300ec9ca36a389aeb510" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.454526 4935 scope.go:117] "RemoveContainer" containerID="6785a03b97344eb642561ee5d12730a7f93de79453bf17b263bd20ae49bbd3b3" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.513743 4935 scope.go:117] "RemoveContainer" containerID="ad1685789306deca0aceea8412449a0dfd2dbc9ebd426662244456257e54f051" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.543577 4935 scope.go:117] "RemoveContainer" containerID="c21841257bca2cda12233fe43bab1edd250a07081b7783d9448defce60d379a8" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.574311 4935 scope.go:117] "RemoveContainer" containerID="7378ebd5875058e6928f9504c6ce673a929f0b593e3c881ba94728592aaa6422" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.606767 4935 scope.go:117] "RemoveContainer" containerID="9b2b88228c11fcf40bd73c44aadf5f4a93619dfc6b49a5c328edcf98f4d93a7e" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.642242 4935 scope.go:117] "RemoveContainer" containerID="83dd3f9b8ccfa6c15a5abedf1cc0dba702eb5c6c024d7e58ef7b5517611fcd66" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.671495 4935 scope.go:117] "RemoveContainer" containerID="24b5a0ee97d3993b21bd1da36e2ba46a94d38c77257bf5f70b85ea35748a294c" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.704058 4935 scope.go:117] "RemoveContainer" containerID="e57b27b19e3967d9be9e7a1b859310794e3bc546a933ee59a48b2cc0ef49ea92" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.734585 4935 scope.go:117] "RemoveContainer" containerID="fc1c1a196142dfa9f9def44b621bd771b1f8fbde0b745b2687d81b122f138212" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.774009 4935 scope.go:117] "RemoveContainer" containerID="886247bb7fb5027803fe13e7b67575d77b9ea5c583be3ae7be247796ac3beed7" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.799849 4935 scope.go:117] "RemoveContainer" containerID="60a3d290ffb9e90b0086750d7b0d89ef70348c9efbf1a8dea1b64dbf0d73f9de" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.834267 4935 scope.go:117] "RemoveContainer" containerID="e86b66c8e04febdd2bd5944c182e0ff7d4dbf92d9fc614f144b437495b40e51c" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.864174 4935 scope.go:117] "RemoveContainer" containerID="77c9c68a700241dead6230cd621003877e33deca266947dce8d521f5290a196a" Oct 05 07:19:53 crc kubenswrapper[4935]: I1005 07:19:53.888599 4935 scope.go:117] "RemoveContainer" containerID="e8f8d95dd578c7c0d462ac9775a7a22a446295ed6782e88a6b08a05a1619462d" Oct 05 07:20:05 crc kubenswrapper[4935]: I1005 07:20:05.777870 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:20:05 crc kubenswrapper[4935]: E1005 07:20:05.779199 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:20:19 crc kubenswrapper[4935]: I1005 07:20:19.777667 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:20:19 crc kubenswrapper[4935]: E1005 07:20:19.779121 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:20:32 crc kubenswrapper[4935]: I1005 07:20:32.779008 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:20:32 crc kubenswrapper[4935]: E1005 07:20:32.779986 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:20:44 crc kubenswrapper[4935]: I1005 07:20:44.777623 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:20:44 crc kubenswrapper[4935]: E1005 07:20:44.778722 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.128869 4935 scope.go:117] "RemoveContainer" containerID="1458d1b7cdd99956b967eb1f9853c2bd78c8f640dd1df44deaebf07a076374dd" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.190085 4935 scope.go:117] "RemoveContainer" containerID="345c0040604dcbfb3be52da65117f237ffd6be39c38f28fb647f162dc6931d3f" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.212176 4935 scope.go:117] "RemoveContainer" containerID="ed56f6247798e566f4409b29c17aa572f76c05d02e825db11fb0cc73b7664132" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.239965 4935 scope.go:117] "RemoveContainer" containerID="1728029339abc70034c54155cfbbf6d41048ac00622ca12313d214f4e7ec9fc9" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.267208 4935 scope.go:117] "RemoveContainer" containerID="f1ba1862b3702dc26abd0d99e354e1a75753aa77005d081bcc2c7164d76ca465" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.294246 4935 scope.go:117] "RemoveContainer" containerID="c9f2b76c7a59a55d95e48d3ed77f8499e7423b0c3445cb1a374ce0c2b6af518e" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.319511 4935 scope.go:117] "RemoveContainer" containerID="b79d9e47fed04b770cc464950c913e995fbd2b7802e83c46a8abf39fd1d704d0" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.370348 4935 scope.go:117] "RemoveContainer" containerID="692ed475fb3453a3e6952fec05183b2dc6f37080fcd7dbbaf53b0d0a2e32e4ac" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.394760 4935 scope.go:117] "RemoveContainer" containerID="da3d1180f3ad1963dd2711515fe79e75473abc8f27f2f472d500b4a81d01765f" Oct 05 07:20:54 crc kubenswrapper[4935]: I1005 07:20:54.413928 4935 scope.go:117] "RemoveContainer" containerID="696aa393f941516e4cfa1b451dd2f3e4440d4dee11149b8365d4707e731a8c95" Oct 05 07:20:58 crc kubenswrapper[4935]: I1005 07:20:58.778227 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:20:58 crc kubenswrapper[4935]: E1005 07:20:58.778870 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:21:09 crc kubenswrapper[4935]: I1005 07:21:09.777146 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:21:09 crc kubenswrapper[4935]: E1005 07:21:09.778228 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:21:24 crc kubenswrapper[4935]: I1005 07:21:24.777635 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:21:24 crc kubenswrapper[4935]: E1005 07:21:24.778878 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:21:37 crc kubenswrapper[4935]: I1005 07:21:37.777809 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:21:37 crc kubenswrapper[4935]: E1005 07:21:37.778853 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:21:48 crc kubenswrapper[4935]: I1005 07:21:48.778033 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:21:48 crc kubenswrapper[4935]: E1005 07:21:48.778632 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:21:54 crc kubenswrapper[4935]: I1005 07:21:54.522291 4935 scope.go:117] "RemoveContainer" containerID="2d66e9ffc7b8a645d646583394f15aa468c427b2fc5f36b059ac5211365b9e21" Oct 05 07:21:54 crc kubenswrapper[4935]: I1005 07:21:54.543436 4935 scope.go:117] "RemoveContainer" containerID="98baaaf4f3d464e0f5da3244138847077c9066de79336a50452f980e1c7db9d4" Oct 05 07:21:59 crc kubenswrapper[4935]: I1005 07:21:59.777362 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:21:59 crc kubenswrapper[4935]: E1005 07:21:59.777870 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:22:12 crc kubenswrapper[4935]: I1005 07:22:12.777492 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:22:12 crc kubenswrapper[4935]: E1005 07:22:12.778649 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:22:24 crc kubenswrapper[4935]: I1005 07:22:24.778551 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:22:24 crc kubenswrapper[4935]: E1005 07:22:24.780019 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:22:39 crc kubenswrapper[4935]: I1005 07:22:39.777610 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:22:39 crc kubenswrapper[4935]: E1005 07:22:39.778371 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:22:50 crc kubenswrapper[4935]: I1005 07:22:50.777151 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:22:51 crc kubenswrapper[4935]: I1005 07:22:51.253840 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"e97848ea56eb00526d9a4d3c4692dae7b079c36a5c1450b0dc33d35cf517999b"} Oct 05 07:25:14 crc kubenswrapper[4935]: I1005 07:25:14.289675 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:25:14 crc kubenswrapper[4935]: I1005 07:25:14.290422 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:25:44 crc kubenswrapper[4935]: I1005 07:25:44.289620 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:25:44 crc kubenswrapper[4935]: I1005 07:25:44.290295 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:26:14 crc kubenswrapper[4935]: I1005 07:26:14.289141 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:26:14 crc kubenswrapper[4935]: I1005 07:26:14.289701 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:26:14 crc kubenswrapper[4935]: I1005 07:26:14.289746 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:26:14 crc kubenswrapper[4935]: I1005 07:26:14.290158 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e97848ea56eb00526d9a4d3c4692dae7b079c36a5c1450b0dc33d35cf517999b"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:26:14 crc kubenswrapper[4935]: I1005 07:26:14.290213 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://e97848ea56eb00526d9a4d3c4692dae7b079c36a5c1450b0dc33d35cf517999b" gracePeriod=600 Oct 05 07:26:15 crc kubenswrapper[4935]: I1005 07:26:15.107685 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="e97848ea56eb00526d9a4d3c4692dae7b079c36a5c1450b0dc33d35cf517999b" exitCode=0 Oct 05 07:26:15 crc kubenswrapper[4935]: I1005 07:26:15.107774 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"e97848ea56eb00526d9a4d3c4692dae7b079c36a5c1450b0dc33d35cf517999b"} Oct 05 07:26:15 crc kubenswrapper[4935]: I1005 07:26:15.108076 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549"} Oct 05 07:26:15 crc kubenswrapper[4935]: I1005 07:26:15.108096 4935 scope.go:117] "RemoveContainer" containerID="fdef5f4f77ab5780ec4d72669806989e0a836322a075306a97f4f2ba77951072" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.325564 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bmm8h"] Oct 05 07:26:47 crc kubenswrapper[4935]: E1005 07:26:47.326490 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerName="extract-utilities" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.326512 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerName="extract-utilities" Oct 05 07:26:47 crc kubenswrapper[4935]: E1005 07:26:47.326530 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerName="registry-server" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.326540 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerName="registry-server" Oct 05 07:26:47 crc kubenswrapper[4935]: E1005 07:26:47.326555 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerName="extract-content" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.326564 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerName="extract-content" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.326825 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="adb043d5-2816-4b9f-82b0-1f89170c0cdc" containerName="registry-server" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.328472 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.352848 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bmm8h"] Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.421730 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9949\" (UniqueName: \"kubernetes.io/projected/2bc3d816-d522-4b0e-b553-d6982bc1d82a-kube-api-access-z9949\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.421827 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-utilities\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.421982 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-catalog-content\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.523209 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9949\" (UniqueName: \"kubernetes.io/projected/2bc3d816-d522-4b0e-b553-d6982bc1d82a-kube-api-access-z9949\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.523294 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-utilities\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.523457 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-catalog-content\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.523757 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-utilities\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.523907 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-catalog-content\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.541434 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9949\" (UniqueName: \"kubernetes.io/projected/2bc3d816-d522-4b0e-b553-d6982bc1d82a-kube-api-access-z9949\") pod \"redhat-operators-bmm8h\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.668320 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:47 crc kubenswrapper[4935]: I1005 07:26:47.889216 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bmm8h"] Oct 05 07:26:48 crc kubenswrapper[4935]: I1005 07:26:48.402879 4935 generic.go:334] "Generic (PLEG): container finished" podID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerID="67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417" exitCode=0 Oct 05 07:26:48 crc kubenswrapper[4935]: I1005 07:26:48.402951 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmm8h" event={"ID":"2bc3d816-d522-4b0e-b553-d6982bc1d82a","Type":"ContainerDied","Data":"67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417"} Oct 05 07:26:48 crc kubenswrapper[4935]: I1005 07:26:48.403219 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmm8h" event={"ID":"2bc3d816-d522-4b0e-b553-d6982bc1d82a","Type":"ContainerStarted","Data":"7e18da42744d1fb7bd2cd2eab3662820ea8d4a2987e7062d5b636d84afe06ede"} Oct 05 07:26:48 crc kubenswrapper[4935]: I1005 07:26:48.405347 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:26:49 crc kubenswrapper[4935]: I1005 07:26:49.414532 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmm8h" event={"ID":"2bc3d816-d522-4b0e-b553-d6982bc1d82a","Type":"ContainerStarted","Data":"225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122"} Oct 05 07:26:50 crc kubenswrapper[4935]: I1005 07:26:50.429291 4935 generic.go:334] "Generic (PLEG): container finished" podID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerID="225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122" exitCode=0 Oct 05 07:26:50 crc kubenswrapper[4935]: I1005 07:26:50.429410 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmm8h" event={"ID":"2bc3d816-d522-4b0e-b553-d6982bc1d82a","Type":"ContainerDied","Data":"225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122"} Oct 05 07:26:51 crc kubenswrapper[4935]: I1005 07:26:51.441012 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmm8h" event={"ID":"2bc3d816-d522-4b0e-b553-d6982bc1d82a","Type":"ContainerStarted","Data":"448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29"} Oct 05 07:26:51 crc kubenswrapper[4935]: I1005 07:26:51.474721 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bmm8h" podStartSLOduration=2.030509904 podStartE2EDuration="4.474695241s" podCreationTimestamp="2025-10-05 07:26:47 +0000 UTC" firstStartedPulling="2025-10-05 07:26:48.405163518 +0000 UTC m=+2042.287789978" lastFinishedPulling="2025-10-05 07:26:50.849348815 +0000 UTC m=+2044.731975315" observedRunningTime="2025-10-05 07:26:51.467277964 +0000 UTC m=+2045.349904444" watchObservedRunningTime="2025-10-05 07:26:51.474695241 +0000 UTC m=+2045.357321741" Oct 05 07:26:57 crc kubenswrapper[4935]: I1005 07:26:57.669361 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:57 crc kubenswrapper[4935]: I1005 07:26:57.669870 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:57 crc kubenswrapper[4935]: I1005 07:26:57.748799 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:58 crc kubenswrapper[4935]: I1005 07:26:58.580830 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:26:58 crc kubenswrapper[4935]: I1005 07:26:58.642272 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bmm8h"] Oct 05 07:27:00 crc kubenswrapper[4935]: I1005 07:27:00.519556 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bmm8h" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerName="registry-server" containerID="cri-o://448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29" gracePeriod=2 Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.066262 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.161440 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-catalog-content\") pod \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.161619 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9949\" (UniqueName: \"kubernetes.io/projected/2bc3d816-d522-4b0e-b553-d6982bc1d82a-kube-api-access-z9949\") pod \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.161779 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-utilities\") pod \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\" (UID: \"2bc3d816-d522-4b0e-b553-d6982bc1d82a\") " Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.163516 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-utilities" (OuterVolumeSpecName: "utilities") pod "2bc3d816-d522-4b0e-b553-d6982bc1d82a" (UID: "2bc3d816-d522-4b0e-b553-d6982bc1d82a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.168779 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bc3d816-d522-4b0e-b553-d6982bc1d82a-kube-api-access-z9949" (OuterVolumeSpecName: "kube-api-access-z9949") pod "2bc3d816-d522-4b0e-b553-d6982bc1d82a" (UID: "2bc3d816-d522-4b0e-b553-d6982bc1d82a"). InnerVolumeSpecName "kube-api-access-z9949". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.264157 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.264228 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9949\" (UniqueName: \"kubernetes.io/projected/2bc3d816-d522-4b0e-b553-d6982bc1d82a-kube-api-access-z9949\") on node \"crc\" DevicePath \"\"" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.265651 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bc3d816-d522-4b0e-b553-d6982bc1d82a" (UID: "2bc3d816-d522-4b0e-b553-d6982bc1d82a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.365191 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bc3d816-d522-4b0e-b553-d6982bc1d82a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.534058 4935 generic.go:334] "Generic (PLEG): container finished" podID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerID="448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29" exitCode=0 Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.534114 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmm8h" event={"ID":"2bc3d816-d522-4b0e-b553-d6982bc1d82a","Type":"ContainerDied","Data":"448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29"} Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.534202 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmm8h" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.534278 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmm8h" event={"ID":"2bc3d816-d522-4b0e-b553-d6982bc1d82a","Type":"ContainerDied","Data":"7e18da42744d1fb7bd2cd2eab3662820ea8d4a2987e7062d5b636d84afe06ede"} Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.534312 4935 scope.go:117] "RemoveContainer" containerID="448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.579220 4935 scope.go:117] "RemoveContainer" containerID="225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.589066 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bmm8h"] Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.594035 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bmm8h"] Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.629692 4935 scope.go:117] "RemoveContainer" containerID="67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.651626 4935 scope.go:117] "RemoveContainer" containerID="448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29" Oct 05 07:27:01 crc kubenswrapper[4935]: E1005 07:27:01.652276 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29\": container with ID starting with 448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29 not found: ID does not exist" containerID="448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.652391 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29"} err="failed to get container status \"448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29\": rpc error: code = NotFound desc = could not find container \"448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29\": container with ID starting with 448b898c254d97723c1249e94cf85cef7b683fc0dff24d60af2a99ec1972ab29 not found: ID does not exist" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.652423 4935 scope.go:117] "RemoveContainer" containerID="225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122" Oct 05 07:27:01 crc kubenswrapper[4935]: E1005 07:27:01.652831 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122\": container with ID starting with 225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122 not found: ID does not exist" containerID="225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.652861 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122"} err="failed to get container status \"225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122\": rpc error: code = NotFound desc = could not find container \"225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122\": container with ID starting with 225482077e95aaada1af14c87c905f1fa261fa9895b8fb17bac74b7c3c66e122 not found: ID does not exist" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.652881 4935 scope.go:117] "RemoveContainer" containerID="67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417" Oct 05 07:27:01 crc kubenswrapper[4935]: E1005 07:27:01.653613 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417\": container with ID starting with 67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417 not found: ID does not exist" containerID="67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417" Oct 05 07:27:01 crc kubenswrapper[4935]: I1005 07:27:01.653661 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417"} err="failed to get container status \"67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417\": rpc error: code = NotFound desc = could not find container \"67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417\": container with ID starting with 67f0c1136b45e47a89272bf3a35d2e3aba4e8835e2fa776983484045279b3417 not found: ID does not exist" Oct 05 07:27:02 crc kubenswrapper[4935]: I1005 07:27:02.786955 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" path="/var/lib/kubelet/pods/2bc3d816-d522-4b0e-b553-d6982bc1d82a/volumes" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.246172 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ffxxk"] Oct 05 07:28:00 crc kubenswrapper[4935]: E1005 07:28:00.247703 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerName="registry-server" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.247730 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerName="registry-server" Oct 05 07:28:00 crc kubenswrapper[4935]: E1005 07:28:00.247783 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerName="extract-content" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.247796 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerName="extract-content" Oct 05 07:28:00 crc kubenswrapper[4935]: E1005 07:28:00.247829 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerName="extract-utilities" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.247843 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerName="extract-utilities" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.248198 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bc3d816-d522-4b0e-b553-d6982bc1d82a" containerName="registry-server" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.251778 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.258744 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ffxxk"] Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.347259 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g46gj\" (UniqueName: \"kubernetes.io/projected/e9367089-2764-4ff4-b914-e7a7f6c47753-kube-api-access-g46gj\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.347647 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-utilities\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.347867 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-catalog-content\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.449587 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-catalog-content\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.449694 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g46gj\" (UniqueName: \"kubernetes.io/projected/e9367089-2764-4ff4-b914-e7a7f6c47753-kube-api-access-g46gj\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.449725 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-utilities\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.450260 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-catalog-content\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.450351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-utilities\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.473398 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g46gj\" (UniqueName: \"kubernetes.io/projected/e9367089-2764-4ff4-b914-e7a7f6c47753-kube-api-access-g46gj\") pod \"certified-operators-ffxxk\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:00 crc kubenswrapper[4935]: I1005 07:28:00.577232 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:01 crc kubenswrapper[4935]: I1005 07:28:01.125867 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ffxxk"] Oct 05 07:28:02 crc kubenswrapper[4935]: I1005 07:28:02.085703 4935 generic.go:334] "Generic (PLEG): container finished" podID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerID="e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78" exitCode=0 Oct 05 07:28:02 crc kubenswrapper[4935]: I1005 07:28:02.085776 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffxxk" event={"ID":"e9367089-2764-4ff4-b914-e7a7f6c47753","Type":"ContainerDied","Data":"e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78"} Oct 05 07:28:02 crc kubenswrapper[4935]: I1005 07:28:02.085816 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffxxk" event={"ID":"e9367089-2764-4ff4-b914-e7a7f6c47753","Type":"ContainerStarted","Data":"fb28a4e024f5f63f69af6089a60387ca5a548b69274342f4af77734e753d3558"} Oct 05 07:28:03 crc kubenswrapper[4935]: I1005 07:28:03.095772 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffxxk" event={"ID":"e9367089-2764-4ff4-b914-e7a7f6c47753","Type":"ContainerStarted","Data":"c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67"} Oct 05 07:28:04 crc kubenswrapper[4935]: I1005 07:28:04.105636 4935 generic.go:334] "Generic (PLEG): container finished" podID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerID="c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67" exitCode=0 Oct 05 07:28:04 crc kubenswrapper[4935]: I1005 07:28:04.105685 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffxxk" event={"ID":"e9367089-2764-4ff4-b914-e7a7f6c47753","Type":"ContainerDied","Data":"c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67"} Oct 05 07:28:05 crc kubenswrapper[4935]: I1005 07:28:05.114922 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffxxk" event={"ID":"e9367089-2764-4ff4-b914-e7a7f6c47753","Type":"ContainerStarted","Data":"d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd"} Oct 05 07:28:05 crc kubenswrapper[4935]: I1005 07:28:05.139146 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ffxxk" podStartSLOduration=2.64498733 podStartE2EDuration="5.139093759s" podCreationTimestamp="2025-10-05 07:28:00 +0000 UTC" firstStartedPulling="2025-10-05 07:28:02.088669702 +0000 UTC m=+2115.971296172" lastFinishedPulling="2025-10-05 07:28:04.582776141 +0000 UTC m=+2118.465402601" observedRunningTime="2025-10-05 07:28:05.133789708 +0000 UTC m=+2119.016416168" watchObservedRunningTime="2025-10-05 07:28:05.139093759 +0000 UTC m=+2119.021720229" Oct 05 07:28:10 crc kubenswrapper[4935]: I1005 07:28:10.578324 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:10 crc kubenswrapper[4935]: I1005 07:28:10.579283 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:10 crc kubenswrapper[4935]: I1005 07:28:10.640703 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:11 crc kubenswrapper[4935]: I1005 07:28:11.243330 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:11 crc kubenswrapper[4935]: I1005 07:28:11.312281 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ffxxk"] Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.182672 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ffxxk" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerName="registry-server" containerID="cri-o://d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd" gracePeriod=2 Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.616849 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.647447 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-utilities\") pod \"e9367089-2764-4ff4-b914-e7a7f6c47753\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.647500 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-catalog-content\") pod \"e9367089-2764-4ff4-b914-e7a7f6c47753\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.647524 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g46gj\" (UniqueName: \"kubernetes.io/projected/e9367089-2764-4ff4-b914-e7a7f6c47753-kube-api-access-g46gj\") pod \"e9367089-2764-4ff4-b914-e7a7f6c47753\" (UID: \"e9367089-2764-4ff4-b914-e7a7f6c47753\") " Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.648322 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-utilities" (OuterVolumeSpecName: "utilities") pod "e9367089-2764-4ff4-b914-e7a7f6c47753" (UID: "e9367089-2764-4ff4-b914-e7a7f6c47753"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.656042 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9367089-2764-4ff4-b914-e7a7f6c47753-kube-api-access-g46gj" (OuterVolumeSpecName: "kube-api-access-g46gj") pod "e9367089-2764-4ff4-b914-e7a7f6c47753" (UID: "e9367089-2764-4ff4-b914-e7a7f6c47753"). InnerVolumeSpecName "kube-api-access-g46gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.749699 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.749742 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g46gj\" (UniqueName: \"kubernetes.io/projected/e9367089-2764-4ff4-b914-e7a7f6c47753-kube-api-access-g46gj\") on node \"crc\" DevicePath \"\"" Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.937624 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9367089-2764-4ff4-b914-e7a7f6c47753" (UID: "e9367089-2764-4ff4-b914-e7a7f6c47753"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:28:13 crc kubenswrapper[4935]: I1005 07:28:13.953941 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9367089-2764-4ff4-b914-e7a7f6c47753-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.196233 4935 generic.go:334] "Generic (PLEG): container finished" podID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerID="d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd" exitCode=0 Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.196339 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffxxk" event={"ID":"e9367089-2764-4ff4-b914-e7a7f6c47753","Type":"ContainerDied","Data":"d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd"} Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.196389 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ffxxk" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.197176 4935 scope.go:117] "RemoveContainer" containerID="d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.199033 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffxxk" event={"ID":"e9367089-2764-4ff4-b914-e7a7f6c47753","Type":"ContainerDied","Data":"fb28a4e024f5f63f69af6089a60387ca5a548b69274342f4af77734e753d3558"} Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.240940 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ffxxk"] Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.243681 4935 scope.go:117] "RemoveContainer" containerID="c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.248697 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ffxxk"] Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.275401 4935 scope.go:117] "RemoveContainer" containerID="e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.289948 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.290005 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.314815 4935 scope.go:117] "RemoveContainer" containerID="d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd" Oct 05 07:28:14 crc kubenswrapper[4935]: E1005 07:28:14.315876 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd\": container with ID starting with d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd not found: ID does not exist" containerID="d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.315964 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd"} err="failed to get container status \"d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd\": rpc error: code = NotFound desc = could not find container \"d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd\": container with ID starting with d62740bc8115ca15cce59dcca5ab41e75486a5aab4b36b1056bbaec18e94ebdd not found: ID does not exist" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.315999 4935 scope.go:117] "RemoveContainer" containerID="c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67" Oct 05 07:28:14 crc kubenswrapper[4935]: E1005 07:28:14.316439 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67\": container with ID starting with c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67 not found: ID does not exist" containerID="c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.316466 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67"} err="failed to get container status \"c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67\": rpc error: code = NotFound desc = could not find container \"c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67\": container with ID starting with c7783528d5623082ae13d736bb1a2c086f58359c643dde69c36b5c8545ccab67 not found: ID does not exist" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.316491 4935 scope.go:117] "RemoveContainer" containerID="e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78" Oct 05 07:28:14 crc kubenswrapper[4935]: E1005 07:28:14.316905 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78\": container with ID starting with e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78 not found: ID does not exist" containerID="e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.316963 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78"} err="failed to get container status \"e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78\": rpc error: code = NotFound desc = could not find container \"e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78\": container with ID starting with e9d56ad2523cb6044673b89d008f4a420610bb170a2bc2dd6457abc475712e78 not found: ID does not exist" Oct 05 07:28:14 crc kubenswrapper[4935]: I1005 07:28:14.787652 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" path="/var/lib/kubelet/pods/e9367089-2764-4ff4-b914-e7a7f6c47753/volumes" Oct 05 07:28:44 crc kubenswrapper[4935]: I1005 07:28:44.289563 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:28:44 crc kubenswrapper[4935]: I1005 07:28:44.290238 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.941887 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xq45d"] Oct 05 07:28:51 crc kubenswrapper[4935]: E1005 07:28:51.943003 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerName="registry-server" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.943022 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerName="registry-server" Oct 05 07:28:51 crc kubenswrapper[4935]: E1005 07:28:51.943036 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerName="extract-utilities" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.943046 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerName="extract-utilities" Oct 05 07:28:51 crc kubenswrapper[4935]: E1005 07:28:51.943062 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerName="extract-content" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.943071 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerName="extract-content" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.943258 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9367089-2764-4ff4-b914-e7a7f6c47753" containerName="registry-server" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.944526 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.952212 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6962\" (UniqueName: \"kubernetes.io/projected/376857c8-d058-4b33-8492-12701a696aab-kube-api-access-h6962\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.952300 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-utilities\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.952348 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-catalog-content\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:51 crc kubenswrapper[4935]: I1005 07:28:51.993997 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xq45d"] Oct 05 07:28:52 crc kubenswrapper[4935]: I1005 07:28:52.054160 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6962\" (UniqueName: \"kubernetes.io/projected/376857c8-d058-4b33-8492-12701a696aab-kube-api-access-h6962\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:52 crc kubenswrapper[4935]: I1005 07:28:52.054247 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-utilities\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:52 crc kubenswrapper[4935]: I1005 07:28:52.054310 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-catalog-content\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:52 crc kubenswrapper[4935]: I1005 07:28:52.055459 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-catalog-content\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:52 crc kubenswrapper[4935]: I1005 07:28:52.056055 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-utilities\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:52 crc kubenswrapper[4935]: I1005 07:28:52.076129 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6962\" (UniqueName: \"kubernetes.io/projected/376857c8-d058-4b33-8492-12701a696aab-kube-api-access-h6962\") pod \"redhat-marketplace-xq45d\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:52 crc kubenswrapper[4935]: I1005 07:28:52.307020 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:28:52 crc kubenswrapper[4935]: I1005 07:28:52.716650 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xq45d"] Oct 05 07:28:53 crc kubenswrapper[4935]: I1005 07:28:53.537487 4935 generic.go:334] "Generic (PLEG): container finished" podID="376857c8-d058-4b33-8492-12701a696aab" containerID="d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603" exitCode=0 Oct 05 07:28:53 crc kubenswrapper[4935]: I1005 07:28:53.537841 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xq45d" event={"ID":"376857c8-d058-4b33-8492-12701a696aab","Type":"ContainerDied","Data":"d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603"} Oct 05 07:28:53 crc kubenswrapper[4935]: I1005 07:28:53.537886 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xq45d" event={"ID":"376857c8-d058-4b33-8492-12701a696aab","Type":"ContainerStarted","Data":"bc12ffe2229a4666baedd8de1f590f864bdf7a31a0eeabaf517940f205f07a08"} Oct 05 07:28:54 crc kubenswrapper[4935]: I1005 07:28:54.548431 4935 generic.go:334] "Generic (PLEG): container finished" podID="376857c8-d058-4b33-8492-12701a696aab" containerID="9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28" exitCode=0 Oct 05 07:28:54 crc kubenswrapper[4935]: I1005 07:28:54.548537 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xq45d" event={"ID":"376857c8-d058-4b33-8492-12701a696aab","Type":"ContainerDied","Data":"9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28"} Oct 05 07:28:55 crc kubenswrapper[4935]: I1005 07:28:55.561656 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xq45d" event={"ID":"376857c8-d058-4b33-8492-12701a696aab","Type":"ContainerStarted","Data":"b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77"} Oct 05 07:28:55 crc kubenswrapper[4935]: I1005 07:28:55.591089 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xq45d" podStartSLOduration=3.169694491 podStartE2EDuration="4.591070121s" podCreationTimestamp="2025-10-05 07:28:51 +0000 UTC" firstStartedPulling="2025-10-05 07:28:53.540179407 +0000 UTC m=+2167.422805867" lastFinishedPulling="2025-10-05 07:28:54.961555037 +0000 UTC m=+2168.844181497" observedRunningTime="2025-10-05 07:28:55.583116521 +0000 UTC m=+2169.465742991" watchObservedRunningTime="2025-10-05 07:28:55.591070121 +0000 UTC m=+2169.473696601" Oct 05 07:29:02 crc kubenswrapper[4935]: I1005 07:29:02.307951 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:29:02 crc kubenswrapper[4935]: I1005 07:29:02.308278 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:29:02 crc kubenswrapper[4935]: I1005 07:29:02.400876 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:29:02 crc kubenswrapper[4935]: I1005 07:29:02.712675 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:29:02 crc kubenswrapper[4935]: I1005 07:29:02.803038 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xq45d"] Oct 05 07:29:04 crc kubenswrapper[4935]: I1005 07:29:04.663130 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xq45d" podUID="376857c8-d058-4b33-8492-12701a696aab" containerName="registry-server" containerID="cri-o://b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77" gracePeriod=2 Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.047545 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.149403 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-catalog-content\") pod \"376857c8-d058-4b33-8492-12701a696aab\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.149474 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6962\" (UniqueName: \"kubernetes.io/projected/376857c8-d058-4b33-8492-12701a696aab-kube-api-access-h6962\") pod \"376857c8-d058-4b33-8492-12701a696aab\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.151120 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-utilities" (OuterVolumeSpecName: "utilities") pod "376857c8-d058-4b33-8492-12701a696aab" (UID: "376857c8-d058-4b33-8492-12701a696aab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.151208 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-utilities\") pod \"376857c8-d058-4b33-8492-12701a696aab\" (UID: \"376857c8-d058-4b33-8492-12701a696aab\") " Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.151490 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.154303 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/376857c8-d058-4b33-8492-12701a696aab-kube-api-access-h6962" (OuterVolumeSpecName: "kube-api-access-h6962") pod "376857c8-d058-4b33-8492-12701a696aab" (UID: "376857c8-d058-4b33-8492-12701a696aab"). InnerVolumeSpecName "kube-api-access-h6962". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.163900 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "376857c8-d058-4b33-8492-12701a696aab" (UID: "376857c8-d058-4b33-8492-12701a696aab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.253118 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376857c8-d058-4b33-8492-12701a696aab-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.253161 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6962\" (UniqueName: \"kubernetes.io/projected/376857c8-d058-4b33-8492-12701a696aab-kube-api-access-h6962\") on node \"crc\" DevicePath \"\"" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.692408 4935 generic.go:334] "Generic (PLEG): container finished" podID="376857c8-d058-4b33-8492-12701a696aab" containerID="b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77" exitCode=0 Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.692457 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xq45d" event={"ID":"376857c8-d058-4b33-8492-12701a696aab","Type":"ContainerDied","Data":"b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77"} Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.692487 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xq45d" event={"ID":"376857c8-d058-4b33-8492-12701a696aab","Type":"ContainerDied","Data":"bc12ffe2229a4666baedd8de1f590f864bdf7a31a0eeabaf517940f205f07a08"} Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.692508 4935 scope.go:117] "RemoveContainer" containerID="b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.692646 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xq45d" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.725915 4935 scope.go:117] "RemoveContainer" containerID="9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.760705 4935 scope.go:117] "RemoveContainer" containerID="d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.819074 4935 scope.go:117] "RemoveContainer" containerID="b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77" Oct 05 07:29:05 crc kubenswrapper[4935]: E1005 07:29:05.820198 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77\": container with ID starting with b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77 not found: ID does not exist" containerID="b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.820238 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77"} err="failed to get container status \"b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77\": rpc error: code = NotFound desc = could not find container \"b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77\": container with ID starting with b1cf31752eadf4b2956bab473d30a492303f13e64f48d4a64ceb533372625d77 not found: ID does not exist" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.820265 4935 scope.go:117] "RemoveContainer" containerID="9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28" Oct 05 07:29:05 crc kubenswrapper[4935]: E1005 07:29:05.827052 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28\": container with ID starting with 9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28 not found: ID does not exist" containerID="9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.827118 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28"} err="failed to get container status \"9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28\": rpc error: code = NotFound desc = could not find container \"9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28\": container with ID starting with 9c1a42651efe9c9ddac44c52dd1e7f9d7f9ccafeb903e059d94930c6b00a5d28 not found: ID does not exist" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.827154 4935 scope.go:117] "RemoveContainer" containerID="d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.827395 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xq45d"] Oct 05 07:29:05 crc kubenswrapper[4935]: E1005 07:29:05.828134 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603\": container with ID starting with d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603 not found: ID does not exist" containerID="d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.828161 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603"} err="failed to get container status \"d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603\": rpc error: code = NotFound desc = could not find container \"d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603\": container with ID starting with d6116c61689bebbe934e9c223e06f05c28e59dbeecfd0a0827b39232b3673603 not found: ID does not exist" Oct 05 07:29:05 crc kubenswrapper[4935]: I1005 07:29:05.832206 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xq45d"] Oct 05 07:29:06 crc kubenswrapper[4935]: I1005 07:29:06.794414 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="376857c8-d058-4b33-8492-12701a696aab" path="/var/lib/kubelet/pods/376857c8-d058-4b33-8492-12701a696aab/volumes" Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.289866 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.290439 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.290491 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.291139 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.291185 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" gracePeriod=600 Oct 05 07:29:14 crc kubenswrapper[4935]: E1005 07:29:14.419617 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.780382 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" exitCode=0 Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.786589 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549"} Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.786966 4935 scope.go:117] "RemoveContainer" containerID="e97848ea56eb00526d9a4d3c4692dae7b079c36a5c1450b0dc33d35cf517999b" Oct 05 07:29:14 crc kubenswrapper[4935]: I1005 07:29:14.787831 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:29:14 crc kubenswrapper[4935]: E1005 07:29:14.788312 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:29:27 crc kubenswrapper[4935]: I1005 07:29:27.777082 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:29:27 crc kubenswrapper[4935]: E1005 07:29:27.778027 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:29:38 crc kubenswrapper[4935]: I1005 07:29:38.777673 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:29:38 crc kubenswrapper[4935]: E1005 07:29:38.778421 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:29:50 crc kubenswrapper[4935]: I1005 07:29:50.778034 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:29:50 crc kubenswrapper[4935]: E1005 07:29:50.778976 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.164683 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6"] Oct 05 07:30:00 crc kubenswrapper[4935]: E1005 07:30:00.165550 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376857c8-d058-4b33-8492-12701a696aab" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.165567 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="376857c8-d058-4b33-8492-12701a696aab" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4935]: E1005 07:30:00.165581 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376857c8-d058-4b33-8492-12701a696aab" containerName="extract-utilities" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.165590 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="376857c8-d058-4b33-8492-12701a696aab" containerName="extract-utilities" Oct 05 07:30:00 crc kubenswrapper[4935]: E1005 07:30:00.165623 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376857c8-d058-4b33-8492-12701a696aab" containerName="extract-content" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.165632 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="376857c8-d058-4b33-8492-12701a696aab" containerName="extract-content" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.165829 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="376857c8-d058-4b33-8492-12701a696aab" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.166393 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.170297 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.173608 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.178045 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6"] Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.212413 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d775t\" (UniqueName: \"kubernetes.io/projected/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-kube-api-access-d775t\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.212475 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-secret-volume\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.212568 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-config-volume\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.314216 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-config-volume\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.314313 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d775t\" (UniqueName: \"kubernetes.io/projected/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-kube-api-access-d775t\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.314351 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-secret-volume\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.315140 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-config-volume\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.332372 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d775t\" (UniqueName: \"kubernetes.io/projected/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-kube-api-access-d775t\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.332705 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-secret-volume\") pod \"collect-profiles-29327490-dgnh6\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.489060 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:00 crc kubenswrapper[4935]: I1005 07:30:00.701261 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6"] Oct 05 07:30:01 crc kubenswrapper[4935]: I1005 07:30:01.190020 4935 generic.go:334] "Generic (PLEG): container finished" podID="03cd4b1b-dcaf-4c59-ba06-b8c11b061473" containerID="24bc98efcf1c1d38bdba0dd62124c501ac653ecf08f63074bb6714380331abcf" exitCode=0 Oct 05 07:30:01 crc kubenswrapper[4935]: I1005 07:30:01.190751 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" event={"ID":"03cd4b1b-dcaf-4c59-ba06-b8c11b061473","Type":"ContainerDied","Data":"24bc98efcf1c1d38bdba0dd62124c501ac653ecf08f63074bb6714380331abcf"} Oct 05 07:30:01 crc kubenswrapper[4935]: I1005 07:30:01.192090 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" event={"ID":"03cd4b1b-dcaf-4c59-ba06-b8c11b061473","Type":"ContainerStarted","Data":"2e141f50824263962c38ee01dfbf4b3dec5abf89d95810cab8ff8a32885f6a13"} Oct 05 07:30:01 crc kubenswrapper[4935]: I1005 07:30:01.778012 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:30:01 crc kubenswrapper[4935]: E1005 07:30:01.778522 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.539613 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.648504 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-secret-volume\") pod \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.648625 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d775t\" (UniqueName: \"kubernetes.io/projected/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-kube-api-access-d775t\") pod \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.648740 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-config-volume\") pod \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\" (UID: \"03cd4b1b-dcaf-4c59-ba06-b8c11b061473\") " Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.649957 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-config-volume" (OuterVolumeSpecName: "config-volume") pod "03cd4b1b-dcaf-4c59-ba06-b8c11b061473" (UID: "03cd4b1b-dcaf-4c59-ba06-b8c11b061473"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.654826 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-kube-api-access-d775t" (OuterVolumeSpecName: "kube-api-access-d775t") pod "03cd4b1b-dcaf-4c59-ba06-b8c11b061473" (UID: "03cd4b1b-dcaf-4c59-ba06-b8c11b061473"). InnerVolumeSpecName "kube-api-access-d775t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.655474 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "03cd4b1b-dcaf-4c59-ba06-b8c11b061473" (UID: "03cd4b1b-dcaf-4c59-ba06-b8c11b061473"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.751012 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.751056 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:30:02 crc kubenswrapper[4935]: I1005 07:30:02.751070 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d775t\" (UniqueName: \"kubernetes.io/projected/03cd4b1b-dcaf-4c59-ba06-b8c11b061473-kube-api-access-d775t\") on node \"crc\" DevicePath \"\"" Oct 05 07:30:03 crc kubenswrapper[4935]: I1005 07:30:03.209507 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" event={"ID":"03cd4b1b-dcaf-4c59-ba06-b8c11b061473","Type":"ContainerDied","Data":"2e141f50824263962c38ee01dfbf4b3dec5abf89d95810cab8ff8a32885f6a13"} Oct 05 07:30:03 crc kubenswrapper[4935]: I1005 07:30:03.209589 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e141f50824263962c38ee01dfbf4b3dec5abf89d95810cab8ff8a32885f6a13" Oct 05 07:30:03 crc kubenswrapper[4935]: I1005 07:30:03.209601 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6" Oct 05 07:30:03 crc kubenswrapper[4935]: I1005 07:30:03.618052 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj"] Oct 05 07:30:03 crc kubenswrapper[4935]: I1005 07:30:03.623271 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-sh9hj"] Oct 05 07:30:04 crc kubenswrapper[4935]: I1005 07:30:04.794805 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce9bae6-1db9-4a69-bef4-1f5da3ea1991" path="/var/lib/kubelet/pods/7ce9bae6-1db9-4a69-bef4-1f5da3ea1991/volumes" Oct 05 07:30:14 crc kubenswrapper[4935]: I1005 07:30:14.778179 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:30:14 crc kubenswrapper[4935]: E1005 07:30:14.779212 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:30:26 crc kubenswrapper[4935]: I1005 07:30:26.785746 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:30:26 crc kubenswrapper[4935]: E1005 07:30:26.787092 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:30:41 crc kubenswrapper[4935]: I1005 07:30:41.777340 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:30:41 crc kubenswrapper[4935]: E1005 07:30:41.777976 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:30:54 crc kubenswrapper[4935]: I1005 07:30:54.778634 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:30:54 crc kubenswrapper[4935]: E1005 07:30:54.780046 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:30:54 crc kubenswrapper[4935]: I1005 07:30:54.819760 4935 scope.go:117] "RemoveContainer" containerID="8d7cc13a8ea0f2e82f746e73d29455e0494bf30064b7deb8315eeb5a46737492" Oct 05 07:31:07 crc kubenswrapper[4935]: I1005 07:31:07.777435 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:31:07 crc kubenswrapper[4935]: E1005 07:31:07.778331 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:31:08 crc kubenswrapper[4935]: I1005 07:31:08.916485 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8zchl"] Oct 05 07:31:08 crc kubenswrapper[4935]: E1005 07:31:08.917219 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03cd4b1b-dcaf-4c59-ba06-b8c11b061473" containerName="collect-profiles" Oct 05 07:31:08 crc kubenswrapper[4935]: I1005 07:31:08.917235 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="03cd4b1b-dcaf-4c59-ba06-b8c11b061473" containerName="collect-profiles" Oct 05 07:31:08 crc kubenswrapper[4935]: I1005 07:31:08.917403 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="03cd4b1b-dcaf-4c59-ba06-b8c11b061473" containerName="collect-profiles" Oct 05 07:31:08 crc kubenswrapper[4935]: I1005 07:31:08.918558 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:08 crc kubenswrapper[4935]: I1005 07:31:08.941947 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8zchl"] Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.099677 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-catalog-content\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.100032 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d62lp\" (UniqueName: \"kubernetes.io/projected/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-kube-api-access-d62lp\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.100084 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-utilities\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.201449 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-catalog-content\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.201573 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d62lp\" (UniqueName: \"kubernetes.io/projected/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-kube-api-access-d62lp\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.201600 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-utilities\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.202235 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-utilities\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.202350 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-catalog-content\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.223915 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d62lp\" (UniqueName: \"kubernetes.io/projected/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-kube-api-access-d62lp\") pod \"community-operators-8zchl\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.256419 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.628248 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8zchl"] Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.881097 4935 generic.go:334] "Generic (PLEG): container finished" podID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerID="df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a" exitCode=0 Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.881179 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zchl" event={"ID":"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5","Type":"ContainerDied","Data":"df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a"} Oct 05 07:31:09 crc kubenswrapper[4935]: I1005 07:31:09.881575 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zchl" event={"ID":"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5","Type":"ContainerStarted","Data":"8d916107d405bb478603a194f36b02c6db8c4b39f7ee1cb6b8e783676d91045b"} Oct 05 07:31:10 crc kubenswrapper[4935]: I1005 07:31:10.901323 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zchl" event={"ID":"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5","Type":"ContainerStarted","Data":"26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c"} Oct 05 07:31:11 crc kubenswrapper[4935]: I1005 07:31:11.915298 4935 generic.go:334] "Generic (PLEG): container finished" podID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerID="26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c" exitCode=0 Oct 05 07:31:11 crc kubenswrapper[4935]: I1005 07:31:11.915386 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zchl" event={"ID":"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5","Type":"ContainerDied","Data":"26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c"} Oct 05 07:31:12 crc kubenswrapper[4935]: I1005 07:31:12.928669 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zchl" event={"ID":"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5","Type":"ContainerStarted","Data":"c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d"} Oct 05 07:31:12 crc kubenswrapper[4935]: I1005 07:31:12.954827 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8zchl" podStartSLOduration=2.399004097 podStartE2EDuration="4.954810478s" podCreationTimestamp="2025-10-05 07:31:08 +0000 UTC" firstStartedPulling="2025-10-05 07:31:09.883249853 +0000 UTC m=+2303.765876353" lastFinishedPulling="2025-10-05 07:31:12.439056274 +0000 UTC m=+2306.321682734" observedRunningTime="2025-10-05 07:31:12.94961266 +0000 UTC m=+2306.832239140" watchObservedRunningTime="2025-10-05 07:31:12.954810478 +0000 UTC m=+2306.837436948" Oct 05 07:31:19 crc kubenswrapper[4935]: I1005 07:31:19.256826 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:19 crc kubenswrapper[4935]: I1005 07:31:19.258149 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:19 crc kubenswrapper[4935]: I1005 07:31:19.324833 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:20 crc kubenswrapper[4935]: I1005 07:31:20.066988 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:20 crc kubenswrapper[4935]: I1005 07:31:20.140528 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8zchl"] Oct 05 07:31:20 crc kubenswrapper[4935]: I1005 07:31:20.778668 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:31:20 crc kubenswrapper[4935]: E1005 07:31:20.779147 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.012001 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8zchl" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerName="registry-server" containerID="cri-o://c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d" gracePeriod=2 Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.537213 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.728532 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d62lp\" (UniqueName: \"kubernetes.io/projected/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-kube-api-access-d62lp\") pod \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.728667 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-catalog-content\") pod \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.728781 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-utilities\") pod \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\" (UID: \"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5\") " Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.730227 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-utilities" (OuterVolumeSpecName: "utilities") pod "f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" (UID: "f5ce2051-3e08-4198-8ba7-79be8b1ac4d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.740210 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-kube-api-access-d62lp" (OuterVolumeSpecName: "kube-api-access-d62lp") pod "f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" (UID: "f5ce2051-3e08-4198-8ba7-79be8b1ac4d5"). InnerVolumeSpecName "kube-api-access-d62lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.796194 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" (UID: "f5ce2051-3e08-4198-8ba7-79be8b1ac4d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.831096 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.831415 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d62lp\" (UniqueName: \"kubernetes.io/projected/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-kube-api-access-d62lp\") on node \"crc\" DevicePath \"\"" Oct 05 07:31:22 crc kubenswrapper[4935]: I1005 07:31:22.831515 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.022733 4935 generic.go:334] "Generic (PLEG): container finished" podID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerID="c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d" exitCode=0 Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.022794 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zchl" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.022817 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zchl" event={"ID":"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5","Type":"ContainerDied","Data":"c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d"} Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.023192 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zchl" event={"ID":"f5ce2051-3e08-4198-8ba7-79be8b1ac4d5","Type":"ContainerDied","Data":"8d916107d405bb478603a194f36b02c6db8c4b39f7ee1cb6b8e783676d91045b"} Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.023212 4935 scope.go:117] "RemoveContainer" containerID="c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.064826 4935 scope.go:117] "RemoveContainer" containerID="26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.078627 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8zchl"] Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.090436 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8zchl"] Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.099262 4935 scope.go:117] "RemoveContainer" containerID="df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.140797 4935 scope.go:117] "RemoveContainer" containerID="c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d" Oct 05 07:31:23 crc kubenswrapper[4935]: E1005 07:31:23.141433 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d\": container with ID starting with c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d not found: ID does not exist" containerID="c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.141489 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d"} err="failed to get container status \"c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d\": rpc error: code = NotFound desc = could not find container \"c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d\": container with ID starting with c0084f705307b6cdfc18cdb7176e36530bc162798280f7b0917260a9b040b92d not found: ID does not exist" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.141523 4935 scope.go:117] "RemoveContainer" containerID="26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c" Oct 05 07:31:23 crc kubenswrapper[4935]: E1005 07:31:23.142375 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c\": container with ID starting with 26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c not found: ID does not exist" containerID="26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.142448 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c"} err="failed to get container status \"26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c\": rpc error: code = NotFound desc = could not find container \"26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c\": container with ID starting with 26a608872e36f4271b2bd7a53beef9362d4ec998cae98d4ff49343650f88157c not found: ID does not exist" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.142477 4935 scope.go:117] "RemoveContainer" containerID="df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a" Oct 05 07:31:23 crc kubenswrapper[4935]: E1005 07:31:23.144732 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a\": container with ID starting with df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a not found: ID does not exist" containerID="df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a" Oct 05 07:31:23 crc kubenswrapper[4935]: I1005 07:31:23.144781 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a"} err="failed to get container status \"df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a\": rpc error: code = NotFound desc = could not find container \"df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a\": container with ID starting with df2b726e2d4118c2df849205efb7668e162615e8d55cfa5fc0a3761b614a0b6a not found: ID does not exist" Oct 05 07:31:24 crc kubenswrapper[4935]: I1005 07:31:24.788769 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" path="/var/lib/kubelet/pods/f5ce2051-3e08-4198-8ba7-79be8b1ac4d5/volumes" Oct 05 07:31:32 crc kubenswrapper[4935]: I1005 07:31:32.777117 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:31:32 crc kubenswrapper[4935]: E1005 07:31:32.777834 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:31:43 crc kubenswrapper[4935]: I1005 07:31:43.778547 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:31:43 crc kubenswrapper[4935]: E1005 07:31:43.779743 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:31:58 crc kubenswrapper[4935]: I1005 07:31:58.778064 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:31:58 crc kubenswrapper[4935]: E1005 07:31:58.781748 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:32:13 crc kubenswrapper[4935]: I1005 07:32:13.777468 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:32:13 crc kubenswrapper[4935]: E1005 07:32:13.778545 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:32:25 crc kubenswrapper[4935]: I1005 07:32:25.777582 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:32:25 crc kubenswrapper[4935]: E1005 07:32:25.781134 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:32:39 crc kubenswrapper[4935]: I1005 07:32:39.777389 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:32:39 crc kubenswrapper[4935]: E1005 07:32:39.778211 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:32:51 crc kubenswrapper[4935]: I1005 07:32:51.776755 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:32:51 crc kubenswrapper[4935]: E1005 07:32:51.777583 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:33:03 crc kubenswrapper[4935]: I1005 07:33:03.777824 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:33:03 crc kubenswrapper[4935]: E1005 07:33:03.779023 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:33:15 crc kubenswrapper[4935]: I1005 07:33:15.777049 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:33:15 crc kubenswrapper[4935]: E1005 07:33:15.777853 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:33:30 crc kubenswrapper[4935]: I1005 07:33:30.777221 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:33:30 crc kubenswrapper[4935]: E1005 07:33:30.777968 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:33:45 crc kubenswrapper[4935]: I1005 07:33:45.777248 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:33:45 crc kubenswrapper[4935]: E1005 07:33:45.777938 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:33:59 crc kubenswrapper[4935]: I1005 07:33:59.776851 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:33:59 crc kubenswrapper[4935]: E1005 07:33:59.777721 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:34:10 crc kubenswrapper[4935]: I1005 07:34:10.777685 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:34:10 crc kubenswrapper[4935]: E1005 07:34:10.778667 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:34:24 crc kubenswrapper[4935]: I1005 07:34:24.777048 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:34:25 crc kubenswrapper[4935]: I1005 07:34:25.749309 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"e04863d21279763d00f51e4e9b6529e884a653604a841dfaa917f4b7526a2040"} Oct 05 07:36:44 crc kubenswrapper[4935]: I1005 07:36:44.289577 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:36:44 crc kubenswrapper[4935]: I1005 07:36:44.290981 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:37:14 crc kubenswrapper[4935]: I1005 07:37:14.289466 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:37:14 crc kubenswrapper[4935]: I1005 07:37:14.290156 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.101750 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jtt42"] Oct 05 07:37:22 crc kubenswrapper[4935]: E1005 07:37:22.105582 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerName="extract-content" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.108793 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerName="extract-content" Oct 05 07:37:22 crc kubenswrapper[4935]: E1005 07:37:22.108842 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerName="extract-utilities" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.108861 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerName="extract-utilities" Oct 05 07:37:22 crc kubenswrapper[4935]: E1005 07:37:22.108977 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerName="registry-server" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.108995 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerName="registry-server" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.109361 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5ce2051-3e08-4198-8ba7-79be8b1ac4d5" containerName="registry-server" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.112145 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.124307 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jtt42"] Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.205182 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-catalog-content\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.205306 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7dt5\" (UniqueName: \"kubernetes.io/projected/82de52f2-6c1e-4b28-ae86-5f4fee72500d-kube-api-access-x7dt5\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.205379 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-utilities\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.307503 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-utilities\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.307830 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-catalog-content\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.308208 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-utilities\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.308312 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-catalog-content\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.308333 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7dt5\" (UniqueName: \"kubernetes.io/projected/82de52f2-6c1e-4b28-ae86-5f4fee72500d-kube-api-access-x7dt5\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.327754 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7dt5\" (UniqueName: \"kubernetes.io/projected/82de52f2-6c1e-4b28-ae86-5f4fee72500d-kube-api-access-x7dt5\") pod \"redhat-operators-jtt42\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.482062 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:22 crc kubenswrapper[4935]: I1005 07:37:22.887280 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jtt42"] Oct 05 07:37:23 crc kubenswrapper[4935]: I1005 07:37:23.246280 4935 generic.go:334] "Generic (PLEG): container finished" podID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerID="c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18" exitCode=0 Oct 05 07:37:23 crc kubenswrapper[4935]: I1005 07:37:23.246341 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtt42" event={"ID":"82de52f2-6c1e-4b28-ae86-5f4fee72500d","Type":"ContainerDied","Data":"c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18"} Oct 05 07:37:23 crc kubenswrapper[4935]: I1005 07:37:23.248025 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtt42" event={"ID":"82de52f2-6c1e-4b28-ae86-5f4fee72500d","Type":"ContainerStarted","Data":"2e3602408cb11d5e7e3c41792aa52f114e2ee9fc5dbb3e0f8751a2a9883a8fdd"} Oct 05 07:37:23 crc kubenswrapper[4935]: I1005 07:37:23.247900 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:37:24 crc kubenswrapper[4935]: I1005 07:37:24.260974 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtt42" event={"ID":"82de52f2-6c1e-4b28-ae86-5f4fee72500d","Type":"ContainerStarted","Data":"98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729"} Oct 05 07:37:25 crc kubenswrapper[4935]: I1005 07:37:25.270238 4935 generic.go:334] "Generic (PLEG): container finished" podID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerID="98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729" exitCode=0 Oct 05 07:37:25 crc kubenswrapper[4935]: I1005 07:37:25.270301 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtt42" event={"ID":"82de52f2-6c1e-4b28-ae86-5f4fee72500d","Type":"ContainerDied","Data":"98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729"} Oct 05 07:37:26 crc kubenswrapper[4935]: I1005 07:37:26.280166 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtt42" event={"ID":"82de52f2-6c1e-4b28-ae86-5f4fee72500d","Type":"ContainerStarted","Data":"79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c"} Oct 05 07:37:32 crc kubenswrapper[4935]: I1005 07:37:32.482970 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:32 crc kubenswrapper[4935]: I1005 07:37:32.483500 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:32 crc kubenswrapper[4935]: I1005 07:37:32.548133 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:32 crc kubenswrapper[4935]: I1005 07:37:32.573994 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jtt42" podStartSLOduration=8.110082871 podStartE2EDuration="10.573964949s" podCreationTimestamp="2025-10-05 07:37:22 +0000 UTC" firstStartedPulling="2025-10-05 07:37:23.247686584 +0000 UTC m=+2677.130313044" lastFinishedPulling="2025-10-05 07:37:25.711568652 +0000 UTC m=+2679.594195122" observedRunningTime="2025-10-05 07:37:26.322473288 +0000 UTC m=+2680.205099778" watchObservedRunningTime="2025-10-05 07:37:32.573964949 +0000 UTC m=+2686.456591449" Oct 05 07:37:33 crc kubenswrapper[4935]: I1005 07:37:33.391738 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:33 crc kubenswrapper[4935]: I1005 07:37:33.438868 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jtt42"] Oct 05 07:37:35 crc kubenswrapper[4935]: I1005 07:37:35.368484 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jtt42" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerName="registry-server" containerID="cri-o://79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c" gracePeriod=2 Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.343945 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.397435 4935 generic.go:334] "Generic (PLEG): container finished" podID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerID="79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c" exitCode=0 Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.397541 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtt42" event={"ID":"82de52f2-6c1e-4b28-ae86-5f4fee72500d","Type":"ContainerDied","Data":"79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c"} Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.397624 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtt42" event={"ID":"82de52f2-6c1e-4b28-ae86-5f4fee72500d","Type":"ContainerDied","Data":"2e3602408cb11d5e7e3c41792aa52f114e2ee9fc5dbb3e0f8751a2a9883a8fdd"} Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.397661 4935 scope.go:117] "RemoveContainer" containerID="79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.397966 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtt42" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.407807 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-utilities\") pod \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.407949 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-catalog-content\") pod \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.408048 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7dt5\" (UniqueName: \"kubernetes.io/projected/82de52f2-6c1e-4b28-ae86-5f4fee72500d-kube-api-access-x7dt5\") pod \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\" (UID: \"82de52f2-6c1e-4b28-ae86-5f4fee72500d\") " Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.409653 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-utilities" (OuterVolumeSpecName: "utilities") pod "82de52f2-6c1e-4b28-ae86-5f4fee72500d" (UID: "82de52f2-6c1e-4b28-ae86-5f4fee72500d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.414616 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82de52f2-6c1e-4b28-ae86-5f4fee72500d-kube-api-access-x7dt5" (OuterVolumeSpecName: "kube-api-access-x7dt5") pod "82de52f2-6c1e-4b28-ae86-5f4fee72500d" (UID: "82de52f2-6c1e-4b28-ae86-5f4fee72500d"). InnerVolumeSpecName "kube-api-access-x7dt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.423974 4935 scope.go:117] "RemoveContainer" containerID="98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.454690 4935 scope.go:117] "RemoveContainer" containerID="c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.481420 4935 scope.go:117] "RemoveContainer" containerID="79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c" Oct 05 07:37:36 crc kubenswrapper[4935]: E1005 07:37:36.481999 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c\": container with ID starting with 79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c not found: ID does not exist" containerID="79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.482055 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c"} err="failed to get container status \"79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c\": rpc error: code = NotFound desc = could not find container \"79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c\": container with ID starting with 79f22921663029de71181ba5b26dcb68d3971712d81384ccff8e03340d0bdf0c not found: ID does not exist" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.482085 4935 scope.go:117] "RemoveContainer" containerID="98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729" Oct 05 07:37:36 crc kubenswrapper[4935]: E1005 07:37:36.482469 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729\": container with ID starting with 98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729 not found: ID does not exist" containerID="98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.482638 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729"} err="failed to get container status \"98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729\": rpc error: code = NotFound desc = could not find container \"98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729\": container with ID starting with 98823f758fbdde0118300376a0c92807ffbcc96d5aaa2ab1a8f003d755a35729 not found: ID does not exist" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.482799 4935 scope.go:117] "RemoveContainer" containerID="c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18" Oct 05 07:37:36 crc kubenswrapper[4935]: E1005 07:37:36.483450 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18\": container with ID starting with c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18 not found: ID does not exist" containerID="c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.483479 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18"} err="failed to get container status \"c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18\": rpc error: code = NotFound desc = could not find container \"c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18\": container with ID starting with c13b9310373fc8490142f2a40e6a274bc8d3ae4a86991fc8f50e5d20c0d73e18 not found: ID does not exist" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.496078 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82de52f2-6c1e-4b28-ae86-5f4fee72500d" (UID: "82de52f2-6c1e-4b28-ae86-5f4fee72500d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.509269 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.509301 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82de52f2-6c1e-4b28-ae86-5f4fee72500d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.509312 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7dt5\" (UniqueName: \"kubernetes.io/projected/82de52f2-6c1e-4b28-ae86-5f4fee72500d-kube-api-access-x7dt5\") on node \"crc\" DevicePath \"\"" Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.737204 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jtt42"] Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.746662 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jtt42"] Oct 05 07:37:36 crc kubenswrapper[4935]: I1005 07:37:36.785981 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" path="/var/lib/kubelet/pods/82de52f2-6c1e-4b28-ae86-5f4fee72500d/volumes" Oct 05 07:37:44 crc kubenswrapper[4935]: I1005 07:37:44.290144 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:37:44 crc kubenswrapper[4935]: I1005 07:37:44.290729 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:37:44 crc kubenswrapper[4935]: I1005 07:37:44.290781 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:37:44 crc kubenswrapper[4935]: I1005 07:37:44.291439 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e04863d21279763d00f51e4e9b6529e884a653604a841dfaa917f4b7526a2040"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:37:44 crc kubenswrapper[4935]: I1005 07:37:44.291486 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://e04863d21279763d00f51e4e9b6529e884a653604a841dfaa917f4b7526a2040" gracePeriod=600 Oct 05 07:37:44 crc kubenswrapper[4935]: I1005 07:37:44.473456 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="e04863d21279763d00f51e4e9b6529e884a653604a841dfaa917f4b7526a2040" exitCode=0 Oct 05 07:37:44 crc kubenswrapper[4935]: I1005 07:37:44.473583 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"e04863d21279763d00f51e4e9b6529e884a653604a841dfaa917f4b7526a2040"} Oct 05 07:37:44 crc kubenswrapper[4935]: I1005 07:37:44.474061 4935 scope.go:117] "RemoveContainer" containerID="8043db3a708cce9733dab886457a895c8d95f4db8475d6588f1422c84a601549" Oct 05 07:37:45 crc kubenswrapper[4935]: I1005 07:37:45.481364 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752"} Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.360456 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ts7gc"] Oct 05 07:39:27 crc kubenswrapper[4935]: E1005 07:39:27.362828 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerName="extract-content" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.363001 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerName="extract-content" Oct 05 07:39:27 crc kubenswrapper[4935]: E1005 07:39:27.363127 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerName="registry-server" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.363228 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerName="registry-server" Oct 05 07:39:27 crc kubenswrapper[4935]: E1005 07:39:27.363335 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerName="extract-utilities" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.363440 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerName="extract-utilities" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.363801 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="82de52f2-6c1e-4b28-ae86-5f4fee72500d" containerName="registry-server" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.365540 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.381498 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ts7gc"] Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.479754 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-utilities\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.479830 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcl54\" (UniqueName: \"kubernetes.io/projected/d275afc5-bf7f-4d75-8d0d-4339e7559528-kube-api-access-dcl54\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.479917 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-catalog-content\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.581169 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-utilities\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.581224 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcl54\" (UniqueName: \"kubernetes.io/projected/d275afc5-bf7f-4d75-8d0d-4339e7559528-kube-api-access-dcl54\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.581278 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-catalog-content\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.581629 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-utilities\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.582481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-catalog-content\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.600064 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcl54\" (UniqueName: \"kubernetes.io/projected/d275afc5-bf7f-4d75-8d0d-4339e7559528-kube-api-access-dcl54\") pod \"redhat-marketplace-ts7gc\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:27 crc kubenswrapper[4935]: I1005 07:39:27.694950 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:28 crc kubenswrapper[4935]: I1005 07:39:28.115962 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ts7gc"] Oct 05 07:39:28 crc kubenswrapper[4935]: I1005 07:39:28.396043 4935 generic.go:334] "Generic (PLEG): container finished" podID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerID="3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665" exitCode=0 Oct 05 07:39:28 crc kubenswrapper[4935]: I1005 07:39:28.396159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ts7gc" event={"ID":"d275afc5-bf7f-4d75-8d0d-4339e7559528","Type":"ContainerDied","Data":"3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665"} Oct 05 07:39:28 crc kubenswrapper[4935]: I1005 07:39:28.396242 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ts7gc" event={"ID":"d275afc5-bf7f-4d75-8d0d-4339e7559528","Type":"ContainerStarted","Data":"fbd081f1e02d2e6ebf57b74040ae17e1dbae202c23feb4f6d84c5e66c7643dd3"} Oct 05 07:39:29 crc kubenswrapper[4935]: I1005 07:39:29.407192 4935 generic.go:334] "Generic (PLEG): container finished" podID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerID="129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da" exitCode=0 Oct 05 07:39:29 crc kubenswrapper[4935]: I1005 07:39:29.407254 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ts7gc" event={"ID":"d275afc5-bf7f-4d75-8d0d-4339e7559528","Type":"ContainerDied","Data":"129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da"} Oct 05 07:39:30 crc kubenswrapper[4935]: I1005 07:39:30.433280 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ts7gc" event={"ID":"d275afc5-bf7f-4d75-8d0d-4339e7559528","Type":"ContainerStarted","Data":"cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306"} Oct 05 07:39:30 crc kubenswrapper[4935]: I1005 07:39:30.462563 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ts7gc" podStartSLOduration=2.046109077 podStartE2EDuration="3.462547547s" podCreationTimestamp="2025-10-05 07:39:27 +0000 UTC" firstStartedPulling="2025-10-05 07:39:28.399347334 +0000 UTC m=+2802.281973824" lastFinishedPulling="2025-10-05 07:39:29.815785834 +0000 UTC m=+2803.698412294" observedRunningTime="2025-10-05 07:39:30.456773343 +0000 UTC m=+2804.339399833" watchObservedRunningTime="2025-10-05 07:39:30.462547547 +0000 UTC m=+2804.345174007" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.163639 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7h7r7"] Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.169809 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.185389 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7h7r7"] Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.259744 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrppb\" (UniqueName: \"kubernetes.io/projected/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-kube-api-access-lrppb\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.259808 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-utilities\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.259847 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-catalog-content\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.361961 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrppb\" (UniqueName: \"kubernetes.io/projected/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-kube-api-access-lrppb\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.362042 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-utilities\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.362098 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-catalog-content\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.362725 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-utilities\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.362965 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-catalog-content\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.398234 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrppb\" (UniqueName: \"kubernetes.io/projected/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-kube-api-access-lrppb\") pod \"certified-operators-7h7r7\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.503912 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:32 crc kubenswrapper[4935]: W1005 07:39:32.804121 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89c1ba6d_4eaf_4443_86ac_bcffbe8ac51c.slice/crio-a849270d5bc3fa99f70b98988ad8d074c6bc655693e606e409da0df52cb0bc12 WatchSource:0}: Error finding container a849270d5bc3fa99f70b98988ad8d074c6bc655693e606e409da0df52cb0bc12: Status 404 returned error can't find the container with id a849270d5bc3fa99f70b98988ad8d074c6bc655693e606e409da0df52cb0bc12 Oct 05 07:39:32 crc kubenswrapper[4935]: I1005 07:39:32.806776 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7h7r7"] Oct 05 07:39:33 crc kubenswrapper[4935]: I1005 07:39:33.469571 4935 generic.go:334] "Generic (PLEG): container finished" podID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerID="539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf" exitCode=0 Oct 05 07:39:33 crc kubenswrapper[4935]: I1005 07:39:33.469636 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h7r7" event={"ID":"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c","Type":"ContainerDied","Data":"539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf"} Oct 05 07:39:33 crc kubenswrapper[4935]: I1005 07:39:33.469704 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h7r7" event={"ID":"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c","Type":"ContainerStarted","Data":"a849270d5bc3fa99f70b98988ad8d074c6bc655693e606e409da0df52cb0bc12"} Oct 05 07:39:34 crc kubenswrapper[4935]: I1005 07:39:34.479241 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h7r7" event={"ID":"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c","Type":"ContainerStarted","Data":"7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0"} Oct 05 07:39:35 crc kubenswrapper[4935]: I1005 07:39:35.509251 4935 generic.go:334] "Generic (PLEG): container finished" podID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerID="7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0" exitCode=0 Oct 05 07:39:35 crc kubenswrapper[4935]: I1005 07:39:35.509318 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h7r7" event={"ID":"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c","Type":"ContainerDied","Data":"7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0"} Oct 05 07:39:36 crc kubenswrapper[4935]: I1005 07:39:36.528955 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h7r7" event={"ID":"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c","Type":"ContainerStarted","Data":"0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51"} Oct 05 07:39:36 crc kubenswrapper[4935]: I1005 07:39:36.554836 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7h7r7" podStartSLOduration=1.8388510390000001 podStartE2EDuration="4.554818595s" podCreationTimestamp="2025-10-05 07:39:32 +0000 UTC" firstStartedPulling="2025-10-05 07:39:33.472158748 +0000 UTC m=+2807.354785228" lastFinishedPulling="2025-10-05 07:39:36.188126324 +0000 UTC m=+2810.070752784" observedRunningTime="2025-10-05 07:39:36.551586538 +0000 UTC m=+2810.434212998" watchObservedRunningTime="2025-10-05 07:39:36.554818595 +0000 UTC m=+2810.437445055" Oct 05 07:39:37 crc kubenswrapper[4935]: I1005 07:39:37.695913 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:37 crc kubenswrapper[4935]: I1005 07:39:37.695991 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:37 crc kubenswrapper[4935]: I1005 07:39:37.756231 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:38 crc kubenswrapper[4935]: I1005 07:39:38.609747 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:41 crc kubenswrapper[4935]: I1005 07:39:41.943878 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ts7gc"] Oct 05 07:39:41 crc kubenswrapper[4935]: I1005 07:39:41.944461 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ts7gc" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerName="registry-server" containerID="cri-o://cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306" gracePeriod=2 Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.347360 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.427976 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcl54\" (UniqueName: \"kubernetes.io/projected/d275afc5-bf7f-4d75-8d0d-4339e7559528-kube-api-access-dcl54\") pod \"d275afc5-bf7f-4d75-8d0d-4339e7559528\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.428158 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-catalog-content\") pod \"d275afc5-bf7f-4d75-8d0d-4339e7559528\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.428250 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-utilities\") pod \"d275afc5-bf7f-4d75-8d0d-4339e7559528\" (UID: \"d275afc5-bf7f-4d75-8d0d-4339e7559528\") " Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.430022 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-utilities" (OuterVolumeSpecName: "utilities") pod "d275afc5-bf7f-4d75-8d0d-4339e7559528" (UID: "d275afc5-bf7f-4d75-8d0d-4339e7559528"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.435132 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d275afc5-bf7f-4d75-8d0d-4339e7559528-kube-api-access-dcl54" (OuterVolumeSpecName: "kube-api-access-dcl54") pod "d275afc5-bf7f-4d75-8d0d-4339e7559528" (UID: "d275afc5-bf7f-4d75-8d0d-4339e7559528"). InnerVolumeSpecName "kube-api-access-dcl54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.447524 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d275afc5-bf7f-4d75-8d0d-4339e7559528" (UID: "d275afc5-bf7f-4d75-8d0d-4339e7559528"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.504066 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.504548 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.531023 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.531081 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d275afc5-bf7f-4d75-8d0d-4339e7559528-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.531104 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcl54\" (UniqueName: \"kubernetes.io/projected/d275afc5-bf7f-4d75-8d0d-4339e7559528-kube-api-access-dcl54\") on node \"crc\" DevicePath \"\"" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.591308 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.594814 4935 generic.go:334] "Generic (PLEG): container finished" podID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerID="cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306" exitCode=0 Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.595023 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ts7gc" event={"ID":"d275afc5-bf7f-4d75-8d0d-4339e7559528","Type":"ContainerDied","Data":"cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306"} Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.595115 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ts7gc" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.595274 4935 scope.go:117] "RemoveContainer" containerID="cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.595175 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ts7gc" event={"ID":"d275afc5-bf7f-4d75-8d0d-4339e7559528","Type":"ContainerDied","Data":"fbd081f1e02d2e6ebf57b74040ae17e1dbae202c23feb4f6d84c5e66c7643dd3"} Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.628073 4935 scope.go:117] "RemoveContainer" containerID="129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.660925 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.663696 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ts7gc"] Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.670582 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ts7gc"] Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.673448 4935 scope.go:117] "RemoveContainer" containerID="3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.704786 4935 scope.go:117] "RemoveContainer" containerID="cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306" Oct 05 07:39:42 crc kubenswrapper[4935]: E1005 07:39:42.705353 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306\": container with ID starting with cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306 not found: ID does not exist" containerID="cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.705389 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306"} err="failed to get container status \"cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306\": rpc error: code = NotFound desc = could not find container \"cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306\": container with ID starting with cb5a5547e8c2ce686f84713338ff0c17b8fda61156c5f45c53f4263e39cc9306 not found: ID does not exist" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.705412 4935 scope.go:117] "RemoveContainer" containerID="129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da" Oct 05 07:39:42 crc kubenswrapper[4935]: E1005 07:39:42.705842 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da\": container with ID starting with 129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da not found: ID does not exist" containerID="129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.705924 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da"} err="failed to get container status \"129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da\": rpc error: code = NotFound desc = could not find container \"129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da\": container with ID starting with 129d02f15e8b8df2545b35f610228e68638e9ba8aff1581a3eccda21d2d891da not found: ID does not exist" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.705964 4935 scope.go:117] "RemoveContainer" containerID="3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665" Oct 05 07:39:42 crc kubenswrapper[4935]: E1005 07:39:42.706284 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665\": container with ID starting with 3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665 not found: ID does not exist" containerID="3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.706311 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665"} err="failed to get container status \"3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665\": rpc error: code = NotFound desc = could not find container \"3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665\": container with ID starting with 3e3d9f9f07283407524313100a63697332606dd96b49f74871f5040e2cebf665 not found: ID does not exist" Oct 05 07:39:42 crc kubenswrapper[4935]: I1005 07:39:42.790776 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" path="/var/lib/kubelet/pods/d275afc5-bf7f-4d75-8d0d-4339e7559528/volumes" Oct 05 07:39:43 crc kubenswrapper[4935]: I1005 07:39:43.141930 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7h7r7"] Oct 05 07:39:44 crc kubenswrapper[4935]: I1005 07:39:44.289992 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:39:44 crc kubenswrapper[4935]: I1005 07:39:44.290318 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:39:44 crc kubenswrapper[4935]: I1005 07:39:44.617765 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7h7r7" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerName="registry-server" containerID="cri-o://0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51" gracePeriod=2 Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.133427 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.276247 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrppb\" (UniqueName: \"kubernetes.io/projected/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-kube-api-access-lrppb\") pod \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.276551 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-catalog-content\") pod \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.276780 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-utilities\") pod \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\" (UID: \"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c\") " Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.277784 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-utilities" (OuterVolumeSpecName: "utilities") pod "89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" (UID: "89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.280989 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-kube-api-access-lrppb" (OuterVolumeSpecName: "kube-api-access-lrppb") pod "89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" (UID: "89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c"). InnerVolumeSpecName "kube-api-access-lrppb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.339385 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" (UID: "89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.378163 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.378198 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrppb\" (UniqueName: \"kubernetes.io/projected/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-kube-api-access-lrppb\") on node \"crc\" DevicePath \"\"" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.378211 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.626632 4935 generic.go:334] "Generic (PLEG): container finished" podID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerID="0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51" exitCode=0 Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.626672 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h7r7" event={"ID":"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c","Type":"ContainerDied","Data":"0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51"} Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.626698 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h7r7" event={"ID":"89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c","Type":"ContainerDied","Data":"a849270d5bc3fa99f70b98988ad8d074c6bc655693e606e409da0df52cb0bc12"} Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.626715 4935 scope.go:117] "RemoveContainer" containerID="0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.626720 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7h7r7" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.654056 4935 scope.go:117] "RemoveContainer" containerID="7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.673653 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7h7r7"] Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.674409 4935 scope.go:117] "RemoveContainer" containerID="539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.683409 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7h7r7"] Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.716785 4935 scope.go:117] "RemoveContainer" containerID="0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51" Oct 05 07:39:45 crc kubenswrapper[4935]: E1005 07:39:45.717219 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51\": container with ID starting with 0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51 not found: ID does not exist" containerID="0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.717372 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51"} err="failed to get container status \"0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51\": rpc error: code = NotFound desc = could not find container \"0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51\": container with ID starting with 0cb326bb5f41a1df6ecf865d020ca162320106286b21d80e24f8a99e8dcf1e51 not found: ID does not exist" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.717534 4935 scope.go:117] "RemoveContainer" containerID="7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0" Oct 05 07:39:45 crc kubenswrapper[4935]: E1005 07:39:45.717981 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0\": container with ID starting with 7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0 not found: ID does not exist" containerID="7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.718016 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0"} err="failed to get container status \"7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0\": rpc error: code = NotFound desc = could not find container \"7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0\": container with ID starting with 7ae7a5b5c665e5d9a5528a36a98ff467da63195ba402275a4547b88741e431e0 not found: ID does not exist" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.718042 4935 scope.go:117] "RemoveContainer" containerID="539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf" Oct 05 07:39:45 crc kubenswrapper[4935]: E1005 07:39:45.718347 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf\": container with ID starting with 539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf not found: ID does not exist" containerID="539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf" Oct 05 07:39:45 crc kubenswrapper[4935]: I1005 07:39:45.718396 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf"} err="failed to get container status \"539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf\": rpc error: code = NotFound desc = could not find container \"539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf\": container with ID starting with 539dbf1e22c10732496639141c91d0ce4df7fdbd0631aa6120edbc9f29021aaf not found: ID does not exist" Oct 05 07:39:46 crc kubenswrapper[4935]: I1005 07:39:46.788861 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" path="/var/lib/kubelet/pods/89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c/volumes" Oct 05 07:40:14 crc kubenswrapper[4935]: I1005 07:40:14.289500 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:40:14 crc kubenswrapper[4935]: I1005 07:40:14.290031 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:40:44 crc kubenswrapper[4935]: I1005 07:40:44.289375 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:40:44 crc kubenswrapper[4935]: I1005 07:40:44.289802 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:40:44 crc kubenswrapper[4935]: I1005 07:40:44.289850 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:40:44 crc kubenswrapper[4935]: I1005 07:40:44.290511 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:40:44 crc kubenswrapper[4935]: I1005 07:40:44.290576 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" gracePeriod=600 Oct 05 07:40:44 crc kubenswrapper[4935]: E1005 07:40:44.410116 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:40:45 crc kubenswrapper[4935]: I1005 07:40:45.168954 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752"} Oct 05 07:40:45 crc kubenswrapper[4935]: I1005 07:40:45.168954 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" exitCode=0 Oct 05 07:40:45 crc kubenswrapper[4935]: I1005 07:40:45.169403 4935 scope.go:117] "RemoveContainer" containerID="e04863d21279763d00f51e4e9b6529e884a653604a841dfaa917f4b7526a2040" Oct 05 07:40:45 crc kubenswrapper[4935]: I1005 07:40:45.171570 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:40:45 crc kubenswrapper[4935]: E1005 07:40:45.172410 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:40:55 crc kubenswrapper[4935]: I1005 07:40:55.776596 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:40:55 crc kubenswrapper[4935]: E1005 07:40:55.777255 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:41:07 crc kubenswrapper[4935]: I1005 07:41:07.776863 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:41:07 crc kubenswrapper[4935]: E1005 07:41:07.777611 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:41:20 crc kubenswrapper[4935]: I1005 07:41:20.777975 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:41:20 crc kubenswrapper[4935]: E1005 07:41:20.779052 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:41:35 crc kubenswrapper[4935]: I1005 07:41:35.777572 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:41:35 crc kubenswrapper[4935]: E1005 07:41:35.778646 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:41:50 crc kubenswrapper[4935]: I1005 07:41:50.777488 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:41:50 crc kubenswrapper[4935]: E1005 07:41:50.778711 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:42:02 crc kubenswrapper[4935]: I1005 07:42:02.784138 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:42:02 crc kubenswrapper[4935]: E1005 07:42:02.785098 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:42:13 crc kubenswrapper[4935]: I1005 07:42:13.777358 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:42:13 crc kubenswrapper[4935]: E1005 07:42:13.778401 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:42:27 crc kubenswrapper[4935]: I1005 07:42:27.778337 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:42:27 crc kubenswrapper[4935]: E1005 07:42:27.779673 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:42:42 crc kubenswrapper[4935]: I1005 07:42:42.777803 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:42:42 crc kubenswrapper[4935]: E1005 07:42:42.779138 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:42:57 crc kubenswrapper[4935]: I1005 07:42:57.778224 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:42:57 crc kubenswrapper[4935]: E1005 07:42:57.779508 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:43:11 crc kubenswrapper[4935]: I1005 07:43:11.777728 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:43:11 crc kubenswrapper[4935]: E1005 07:43:11.781057 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:43:25 crc kubenswrapper[4935]: I1005 07:43:25.777457 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:43:25 crc kubenswrapper[4935]: E1005 07:43:25.778669 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:43:37 crc kubenswrapper[4935]: I1005 07:43:37.777687 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:43:37 crc kubenswrapper[4935]: E1005 07:43:37.779250 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:43:50 crc kubenswrapper[4935]: I1005 07:43:50.777604 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:43:50 crc kubenswrapper[4935]: E1005 07:43:50.778917 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:44:01 crc kubenswrapper[4935]: I1005 07:44:01.777533 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:44:01 crc kubenswrapper[4935]: E1005 07:44:01.778354 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:44:12 crc kubenswrapper[4935]: I1005 07:44:12.778047 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:44:12 crc kubenswrapper[4935]: E1005 07:44:12.779426 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:44:26 crc kubenswrapper[4935]: I1005 07:44:26.789655 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:44:26 crc kubenswrapper[4935]: E1005 07:44:26.792841 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:44:40 crc kubenswrapper[4935]: I1005 07:44:40.777800 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:44:40 crc kubenswrapper[4935]: E1005 07:44:40.780276 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:44:53 crc kubenswrapper[4935]: I1005 07:44:53.777631 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:44:53 crc kubenswrapper[4935]: E1005 07:44:53.778847 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.204946 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw"] Oct 05 07:45:00 crc kubenswrapper[4935]: E1005 07:45:00.205942 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerName="extract-utilities" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.205960 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerName="extract-utilities" Oct 05 07:45:00 crc kubenswrapper[4935]: E1005 07:45:00.205977 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerName="extract-utilities" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.205984 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerName="extract-utilities" Oct 05 07:45:00 crc kubenswrapper[4935]: E1005 07:45:00.205992 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerName="extract-content" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.206000 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerName="extract-content" Oct 05 07:45:00 crc kubenswrapper[4935]: E1005 07:45:00.206011 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.206018 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4935]: E1005 07:45:00.206038 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.206045 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4935]: E1005 07:45:00.206057 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerName="extract-content" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.206063 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerName="extract-content" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.206224 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d275afc5-bf7f-4d75-8d0d-4339e7559528" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.206242 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="89c1ba6d-4eaf-4443-86ac-bcffbe8ac51c" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.206831 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.211293 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.218434 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.220320 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw"] Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.319707 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-secret-volume\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.320052 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n5xw\" (UniqueName: \"kubernetes.io/projected/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-kube-api-access-2n5xw\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.320196 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-config-volume\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.421713 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-secret-volume\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.421783 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n5xw\" (UniqueName: \"kubernetes.io/projected/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-kube-api-access-2n5xw\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.421829 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-config-volume\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.422978 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-config-volume\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.431369 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-secret-volume\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.440189 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n5xw\" (UniqueName: \"kubernetes.io/projected/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-kube-api-access-2n5xw\") pod \"collect-profiles-29327505-sxrpw\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.538688 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:00 crc kubenswrapper[4935]: I1005 07:45:00.943148 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw"] Oct 05 07:45:01 crc kubenswrapper[4935]: I1005 07:45:01.506623 4935 generic.go:334] "Generic (PLEG): container finished" podID="7d4f606e-3b56-46fa-b13c-0ca4a9c623e1" containerID="71fca8c49f608da923ef71e1b6d883361ba977867f3788d47af960feef5fb48a" exitCode=0 Oct 05 07:45:01 crc kubenswrapper[4935]: I1005 07:45:01.506687 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" event={"ID":"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1","Type":"ContainerDied","Data":"71fca8c49f608da923ef71e1b6d883361ba977867f3788d47af960feef5fb48a"} Oct 05 07:45:01 crc kubenswrapper[4935]: I1005 07:45:01.506938 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" event={"ID":"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1","Type":"ContainerStarted","Data":"2b81c7ed0bce802df4b64617afca9c8e614d017180103b1706b79a1321ff9ed5"} Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.806577 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.855137 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-secret-volume\") pod \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.855174 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n5xw\" (UniqueName: \"kubernetes.io/projected/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-kube-api-access-2n5xw\") pod \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.855277 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-config-volume\") pod \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\" (UID: \"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1\") " Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.856597 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-config-volume" (OuterVolumeSpecName: "config-volume") pod "7d4f606e-3b56-46fa-b13c-0ca4a9c623e1" (UID: "7d4f606e-3b56-46fa-b13c-0ca4a9c623e1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.868287 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7d4f606e-3b56-46fa-b13c-0ca4a9c623e1" (UID: "7d4f606e-3b56-46fa-b13c-0ca4a9c623e1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.868381 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-kube-api-access-2n5xw" (OuterVolumeSpecName: "kube-api-access-2n5xw") pod "7d4f606e-3b56-46fa-b13c-0ca4a9c623e1" (UID: "7d4f606e-3b56-46fa-b13c-0ca4a9c623e1"). InnerVolumeSpecName "kube-api-access-2n5xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.956862 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.957131 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n5xw\" (UniqueName: \"kubernetes.io/projected/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-kube-api-access-2n5xw\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:02 crc kubenswrapper[4935]: I1005 07:45:02.957205 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:03 crc kubenswrapper[4935]: I1005 07:45:03.525839 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" event={"ID":"7d4f606e-3b56-46fa-b13c-0ca4a9c623e1","Type":"ContainerDied","Data":"2b81c7ed0bce802df4b64617afca9c8e614d017180103b1706b79a1321ff9ed5"} Oct 05 07:45:03 crc kubenswrapper[4935]: I1005 07:45:03.525938 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw" Oct 05 07:45:03 crc kubenswrapper[4935]: I1005 07:45:03.525973 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b81c7ed0bce802df4b64617afca9c8e614d017180103b1706b79a1321ff9ed5" Oct 05 07:45:03 crc kubenswrapper[4935]: I1005 07:45:03.886964 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q"] Oct 05 07:45:03 crc kubenswrapper[4935]: I1005 07:45:03.893148 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-4km5q"] Oct 05 07:45:04 crc kubenswrapper[4935]: I1005 07:45:04.793543 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff3c26d6-6c89-4636-a105-33885e9a9de4" path="/var/lib/kubelet/pods/ff3c26d6-6c89-4636-a105-33885e9a9de4/volumes" Oct 05 07:45:06 crc kubenswrapper[4935]: I1005 07:45:06.782746 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:45:06 crc kubenswrapper[4935]: E1005 07:45:06.783020 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:45:18 crc kubenswrapper[4935]: I1005 07:45:18.777454 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:45:18 crc kubenswrapper[4935]: E1005 07:45:18.778725 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:45:31 crc kubenswrapper[4935]: I1005 07:45:31.778115 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:45:31 crc kubenswrapper[4935]: E1005 07:45:31.779385 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:45:44 crc kubenswrapper[4935]: I1005 07:45:44.778199 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:45:45 crc kubenswrapper[4935]: I1005 07:45:45.881557 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"2142111d863206494d1c9e2a850cd11e2b5a731b9df5e3303b181065609022b4"} Oct 05 07:45:55 crc kubenswrapper[4935]: I1005 07:45:55.175184 4935 scope.go:117] "RemoveContainer" containerID="8203d88f3096ded755782c14c1ae97226db565f9e28b7cb65b400adb68d28f69" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.209188 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qd7r7"] Oct 05 07:47:02 crc kubenswrapper[4935]: E1005 07:47:02.210390 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4f606e-3b56-46fa-b13c-0ca4a9c623e1" containerName="collect-profiles" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.210413 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4f606e-3b56-46fa-b13c-0ca4a9c623e1" containerName="collect-profiles" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.210692 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4f606e-3b56-46fa-b13c-0ca4a9c623e1" containerName="collect-profiles" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.212481 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.223270 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qd7r7"] Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.346162 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-utilities\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.346267 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-catalog-content\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.346319 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8clzm\" (UniqueName: \"kubernetes.io/projected/0e85d87f-6e1b-4712-9889-e0caee8cc678-kube-api-access-8clzm\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.448672 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-catalog-content\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.448734 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8clzm\" (UniqueName: \"kubernetes.io/projected/0e85d87f-6e1b-4712-9889-e0caee8cc678-kube-api-access-8clzm\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.448818 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-utilities\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.449370 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-utilities\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.449366 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-catalog-content\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.472771 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8clzm\" (UniqueName: \"kubernetes.io/projected/0e85d87f-6e1b-4712-9889-e0caee8cc678-kube-api-access-8clzm\") pod \"community-operators-qd7r7\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:02 crc kubenswrapper[4935]: I1005 07:47:02.550484 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:03 crc kubenswrapper[4935]: I1005 07:47:03.072385 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qd7r7"] Oct 05 07:47:03 crc kubenswrapper[4935]: W1005 07:47:03.075127 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e85d87f_6e1b_4712_9889_e0caee8cc678.slice/crio-f3231c57eb32c77f70d05c37f3e2e37136848bf2838cc2a5619feff7f2224369 WatchSource:0}: Error finding container f3231c57eb32c77f70d05c37f3e2e37136848bf2838cc2a5619feff7f2224369: Status 404 returned error can't find the container with id f3231c57eb32c77f70d05c37f3e2e37136848bf2838cc2a5619feff7f2224369 Oct 05 07:47:03 crc kubenswrapper[4935]: I1005 07:47:03.591549 4935 generic.go:334] "Generic (PLEG): container finished" podID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerID="2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde" exitCode=0 Oct 05 07:47:03 crc kubenswrapper[4935]: I1005 07:47:03.591807 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd7r7" event={"ID":"0e85d87f-6e1b-4712-9889-e0caee8cc678","Type":"ContainerDied","Data":"2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde"} Oct 05 07:47:03 crc kubenswrapper[4935]: I1005 07:47:03.591835 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd7r7" event={"ID":"0e85d87f-6e1b-4712-9889-e0caee8cc678","Type":"ContainerStarted","Data":"f3231c57eb32c77f70d05c37f3e2e37136848bf2838cc2a5619feff7f2224369"} Oct 05 07:47:03 crc kubenswrapper[4935]: I1005 07:47:03.595128 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:47:04 crc kubenswrapper[4935]: I1005 07:47:04.604949 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd7r7" event={"ID":"0e85d87f-6e1b-4712-9889-e0caee8cc678","Type":"ContainerStarted","Data":"98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d"} Oct 05 07:47:05 crc kubenswrapper[4935]: I1005 07:47:05.620387 4935 generic.go:334] "Generic (PLEG): container finished" podID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerID="98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d" exitCode=0 Oct 05 07:47:05 crc kubenswrapper[4935]: I1005 07:47:05.620479 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd7r7" event={"ID":"0e85d87f-6e1b-4712-9889-e0caee8cc678","Type":"ContainerDied","Data":"98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d"} Oct 05 07:47:06 crc kubenswrapper[4935]: I1005 07:47:06.634742 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd7r7" event={"ID":"0e85d87f-6e1b-4712-9889-e0caee8cc678","Type":"ContainerStarted","Data":"74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550"} Oct 05 07:47:06 crc kubenswrapper[4935]: I1005 07:47:06.676727 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qd7r7" podStartSLOduration=2.279186456 podStartE2EDuration="4.676699139s" podCreationTimestamp="2025-10-05 07:47:02 +0000 UTC" firstStartedPulling="2025-10-05 07:47:03.594707511 +0000 UTC m=+3257.477334011" lastFinishedPulling="2025-10-05 07:47:05.992220234 +0000 UTC m=+3259.874846694" observedRunningTime="2025-10-05 07:47:06.669745104 +0000 UTC m=+3260.552371624" watchObservedRunningTime="2025-10-05 07:47:06.676699139 +0000 UTC m=+3260.559325639" Oct 05 07:47:12 crc kubenswrapper[4935]: I1005 07:47:12.550735 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:12 crc kubenswrapper[4935]: I1005 07:47:12.552321 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:12 crc kubenswrapper[4935]: I1005 07:47:12.630692 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:12 crc kubenswrapper[4935]: I1005 07:47:12.764515 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:12 crc kubenswrapper[4935]: I1005 07:47:12.876571 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qd7r7"] Oct 05 07:47:14 crc kubenswrapper[4935]: I1005 07:47:14.719257 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qd7r7" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerName="registry-server" containerID="cri-o://74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550" gracePeriod=2 Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.184656 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.347268 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-utilities\") pod \"0e85d87f-6e1b-4712-9889-e0caee8cc678\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.347382 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8clzm\" (UniqueName: \"kubernetes.io/projected/0e85d87f-6e1b-4712-9889-e0caee8cc678-kube-api-access-8clzm\") pod \"0e85d87f-6e1b-4712-9889-e0caee8cc678\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.347450 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-catalog-content\") pod \"0e85d87f-6e1b-4712-9889-e0caee8cc678\" (UID: \"0e85d87f-6e1b-4712-9889-e0caee8cc678\") " Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.348969 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-utilities" (OuterVolumeSpecName: "utilities") pod "0e85d87f-6e1b-4712-9889-e0caee8cc678" (UID: "0e85d87f-6e1b-4712-9889-e0caee8cc678"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.356309 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e85d87f-6e1b-4712-9889-e0caee8cc678-kube-api-access-8clzm" (OuterVolumeSpecName: "kube-api-access-8clzm") pod "0e85d87f-6e1b-4712-9889-e0caee8cc678" (UID: "0e85d87f-6e1b-4712-9889-e0caee8cc678"). InnerVolumeSpecName "kube-api-access-8clzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.410205 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e85d87f-6e1b-4712-9889-e0caee8cc678" (UID: "0e85d87f-6e1b-4712-9889-e0caee8cc678"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.449725 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.449786 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85d87f-6e1b-4712-9889-e0caee8cc678-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.449807 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8clzm\" (UniqueName: \"kubernetes.io/projected/0e85d87f-6e1b-4712-9889-e0caee8cc678-kube-api-access-8clzm\") on node \"crc\" DevicePath \"\"" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.733291 4935 generic.go:334] "Generic (PLEG): container finished" podID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerID="74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550" exitCode=0 Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.733349 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd7r7" event={"ID":"0e85d87f-6e1b-4712-9889-e0caee8cc678","Type":"ContainerDied","Data":"74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550"} Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.733424 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qd7r7" event={"ID":"0e85d87f-6e1b-4712-9889-e0caee8cc678","Type":"ContainerDied","Data":"f3231c57eb32c77f70d05c37f3e2e37136848bf2838cc2a5619feff7f2224369"} Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.733430 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qd7r7" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.733457 4935 scope.go:117] "RemoveContainer" containerID="74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.763333 4935 scope.go:117] "RemoveContainer" containerID="98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.797129 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qd7r7"] Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.799251 4935 scope.go:117] "RemoveContainer" containerID="2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.805284 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qd7r7"] Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.835877 4935 scope.go:117] "RemoveContainer" containerID="74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550" Oct 05 07:47:15 crc kubenswrapper[4935]: E1005 07:47:15.836362 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550\": container with ID starting with 74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550 not found: ID does not exist" containerID="74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.836426 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550"} err="failed to get container status \"74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550\": rpc error: code = NotFound desc = could not find container \"74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550\": container with ID starting with 74acb76522405c169eadb36a8f0a22e9ac51ab2402a4baa233247f62ab9ca550 not found: ID does not exist" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.836453 4935 scope.go:117] "RemoveContainer" containerID="98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d" Oct 05 07:47:15 crc kubenswrapper[4935]: E1005 07:47:15.836839 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d\": container with ID starting with 98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d not found: ID does not exist" containerID="98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.836882 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d"} err="failed to get container status \"98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d\": rpc error: code = NotFound desc = could not find container \"98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d\": container with ID starting with 98f680508ea190035789d7068f0a79adb8c5315dc5fc400f926b8ddf20cccb6d not found: ID does not exist" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.836923 4935 scope.go:117] "RemoveContainer" containerID="2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde" Oct 05 07:47:15 crc kubenswrapper[4935]: E1005 07:47:15.837242 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde\": container with ID starting with 2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde not found: ID does not exist" containerID="2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde" Oct 05 07:47:15 crc kubenswrapper[4935]: I1005 07:47:15.837272 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde"} err="failed to get container status \"2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde\": rpc error: code = NotFound desc = could not find container \"2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde\": container with ID starting with 2a72cb19a73a82217d515bdeee5ad2c6b8ca5fcadfd5b777ef519ca6700afdde not found: ID does not exist" Oct 05 07:47:16 crc kubenswrapper[4935]: I1005 07:47:16.790522 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" path="/var/lib/kubelet/pods/0e85d87f-6e1b-4712-9889-e0caee8cc678/volumes" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.034750 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pbgbj"] Oct 05 07:47:34 crc kubenswrapper[4935]: E1005 07:47:34.041874 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerName="extract-content" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.042012 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerName="extract-content" Oct 05 07:47:34 crc kubenswrapper[4935]: E1005 07:47:34.042079 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerName="extract-utilities" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.042103 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerName="extract-utilities" Oct 05 07:47:34 crc kubenswrapper[4935]: E1005 07:47:34.042149 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerName="registry-server" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.042168 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerName="registry-server" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.042592 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e85d87f-6e1b-4712-9889-e0caee8cc678" containerName="registry-server" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.045417 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.059286 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pbgbj"] Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.134061 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-utilities\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.134353 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-catalog-content\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.134402 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxhsd\" (UniqueName: \"kubernetes.io/projected/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-kube-api-access-vxhsd\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.235761 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-catalog-content\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.235840 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxhsd\" (UniqueName: \"kubernetes.io/projected/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-kube-api-access-vxhsd\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.235924 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-utilities\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.236586 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-catalog-content\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.236604 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-utilities\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.270740 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxhsd\" (UniqueName: \"kubernetes.io/projected/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-kube-api-access-vxhsd\") pod \"redhat-operators-pbgbj\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.384094 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.876032 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pbgbj"] Oct 05 07:47:34 crc kubenswrapper[4935]: I1005 07:47:34.913362 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbgbj" event={"ID":"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1","Type":"ContainerStarted","Data":"c7469a093ccf7fc0379705370f315f4153daa95db6d552a8c93e434e79f8c32a"} Oct 05 07:47:35 crc kubenswrapper[4935]: I1005 07:47:35.924135 4935 generic.go:334] "Generic (PLEG): container finished" podID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerID="ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182" exitCode=0 Oct 05 07:47:35 crc kubenswrapper[4935]: I1005 07:47:35.924233 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbgbj" event={"ID":"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1","Type":"ContainerDied","Data":"ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182"} Oct 05 07:47:36 crc kubenswrapper[4935]: I1005 07:47:36.936000 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbgbj" event={"ID":"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1","Type":"ContainerStarted","Data":"da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287"} Oct 05 07:47:37 crc kubenswrapper[4935]: I1005 07:47:37.946799 4935 generic.go:334] "Generic (PLEG): container finished" podID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerID="da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287" exitCode=0 Oct 05 07:47:37 crc kubenswrapper[4935]: I1005 07:47:37.946977 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbgbj" event={"ID":"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1","Type":"ContainerDied","Data":"da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287"} Oct 05 07:47:38 crc kubenswrapper[4935]: I1005 07:47:38.978426 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbgbj" event={"ID":"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1","Type":"ContainerStarted","Data":"30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a"} Oct 05 07:47:39 crc kubenswrapper[4935]: I1005 07:47:39.008548 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pbgbj" podStartSLOduration=2.449140341 podStartE2EDuration="5.008514403s" podCreationTimestamp="2025-10-05 07:47:34 +0000 UTC" firstStartedPulling="2025-10-05 07:47:35.928397053 +0000 UTC m=+3289.811023553" lastFinishedPulling="2025-10-05 07:47:38.487771145 +0000 UTC m=+3292.370397615" observedRunningTime="2025-10-05 07:47:39.002849552 +0000 UTC m=+3292.885476032" watchObservedRunningTime="2025-10-05 07:47:39.008514403 +0000 UTC m=+3292.891140893" Oct 05 07:47:44 crc kubenswrapper[4935]: I1005 07:47:44.289679 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:47:44 crc kubenswrapper[4935]: I1005 07:47:44.290200 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:47:44 crc kubenswrapper[4935]: I1005 07:47:44.384352 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:44 crc kubenswrapper[4935]: I1005 07:47:44.384439 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:45 crc kubenswrapper[4935]: I1005 07:47:45.460742 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pbgbj" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="registry-server" probeResult="failure" output=< Oct 05 07:47:45 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 07:47:45 crc kubenswrapper[4935]: > Oct 05 07:47:54 crc kubenswrapper[4935]: I1005 07:47:54.464396 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:54 crc kubenswrapper[4935]: I1005 07:47:54.548007 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:54 crc kubenswrapper[4935]: I1005 07:47:54.711416 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pbgbj"] Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.153105 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pbgbj" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="registry-server" containerID="cri-o://30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a" gracePeriod=2 Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.583426 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.721507 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-catalog-content\") pod \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.721581 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxhsd\" (UniqueName: \"kubernetes.io/projected/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-kube-api-access-vxhsd\") pod \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.721618 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-utilities\") pod \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\" (UID: \"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1\") " Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.722960 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-utilities" (OuterVolumeSpecName: "utilities") pod "0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" (UID: "0d8f18aa-57a1-4a86-bb30-f4b7511c31e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.730251 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-kube-api-access-vxhsd" (OuterVolumeSpecName: "kube-api-access-vxhsd") pod "0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" (UID: "0d8f18aa-57a1-4a86-bb30-f4b7511c31e1"). InnerVolumeSpecName "kube-api-access-vxhsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.823087 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxhsd\" (UniqueName: \"kubernetes.io/projected/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-kube-api-access-vxhsd\") on node \"crc\" DevicePath \"\"" Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.823138 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.828474 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" (UID: "0d8f18aa-57a1-4a86-bb30-f4b7511c31e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:47:56 crc kubenswrapper[4935]: I1005 07:47:56.924546 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.164659 4935 generic.go:334] "Generic (PLEG): container finished" podID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerID="30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a" exitCode=0 Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.164709 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbgbj" event={"ID":"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1","Type":"ContainerDied","Data":"30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a"} Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.164750 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbgbj" event={"ID":"0d8f18aa-57a1-4a86-bb30-f4b7511c31e1","Type":"ContainerDied","Data":"c7469a093ccf7fc0379705370f315f4153daa95db6d552a8c93e434e79f8c32a"} Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.164767 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pbgbj" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.164776 4935 scope.go:117] "RemoveContainer" containerID="30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.183715 4935 scope.go:117] "RemoveContainer" containerID="da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.207252 4935 scope.go:117] "RemoveContainer" containerID="ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.215479 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pbgbj"] Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.222685 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pbgbj"] Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.240593 4935 scope.go:117] "RemoveContainer" containerID="30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a" Oct 05 07:47:57 crc kubenswrapper[4935]: E1005 07:47:57.241060 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a\": container with ID starting with 30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a not found: ID does not exist" containerID="30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.241112 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a"} err="failed to get container status \"30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a\": rpc error: code = NotFound desc = could not find container \"30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a\": container with ID starting with 30847537b82dcf0dfadf372d21692440081a823865db1060b5319e4136ee3a1a not found: ID does not exist" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.241132 4935 scope.go:117] "RemoveContainer" containerID="da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287" Oct 05 07:47:57 crc kubenswrapper[4935]: E1005 07:47:57.241527 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287\": container with ID starting with da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287 not found: ID does not exist" containerID="da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.241755 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287"} err="failed to get container status \"da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287\": rpc error: code = NotFound desc = could not find container \"da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287\": container with ID starting with da8b353fb876194b74d5609f6e41c99accca1f7a8ca013987a8bc368a8d67287 not found: ID does not exist" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.241780 4935 scope.go:117] "RemoveContainer" containerID="ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182" Oct 05 07:47:57 crc kubenswrapper[4935]: E1005 07:47:57.242167 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182\": container with ID starting with ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182 not found: ID does not exist" containerID="ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182" Oct 05 07:47:57 crc kubenswrapper[4935]: I1005 07:47:57.242322 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182"} err="failed to get container status \"ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182\": rpc error: code = NotFound desc = could not find container \"ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182\": container with ID starting with ade79fc2ac9aaca2ba01c7deb0df451161548fbd6ddfad83bb98252f85491182 not found: ID does not exist" Oct 05 07:47:58 crc kubenswrapper[4935]: I1005 07:47:58.793877 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" path="/var/lib/kubelet/pods/0d8f18aa-57a1-4a86-bb30-f4b7511c31e1/volumes" Oct 05 07:48:14 crc kubenswrapper[4935]: I1005 07:48:14.289128 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:48:14 crc kubenswrapper[4935]: I1005 07:48:14.289723 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:48:44 crc kubenswrapper[4935]: I1005 07:48:44.289856 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:48:44 crc kubenswrapper[4935]: I1005 07:48:44.290771 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:48:44 crc kubenswrapper[4935]: I1005 07:48:44.290993 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:48:44 crc kubenswrapper[4935]: I1005 07:48:44.292107 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2142111d863206494d1c9e2a850cd11e2b5a731b9df5e3303b181065609022b4"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:48:44 crc kubenswrapper[4935]: I1005 07:48:44.292252 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://2142111d863206494d1c9e2a850cd11e2b5a731b9df5e3303b181065609022b4" gracePeriod=600 Oct 05 07:48:44 crc kubenswrapper[4935]: I1005 07:48:44.574959 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="2142111d863206494d1c9e2a850cd11e2b5a731b9df5e3303b181065609022b4" exitCode=0 Oct 05 07:48:44 crc kubenswrapper[4935]: I1005 07:48:44.575026 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"2142111d863206494d1c9e2a850cd11e2b5a731b9df5e3303b181065609022b4"} Oct 05 07:48:44 crc kubenswrapper[4935]: I1005 07:48:44.575342 4935 scope.go:117] "RemoveContainer" containerID="71fc192a683162601e78aac360ac9df4bff84ae8ad972d49b89998b9eee5d752" Oct 05 07:48:45 crc kubenswrapper[4935]: I1005 07:48:45.584206 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4"} Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.505821 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4jkpk"] Oct 05 07:50:42 crc kubenswrapper[4935]: E1005 07:50:42.506982 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="extract-content" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.507006 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="extract-content" Oct 05 07:50:42 crc kubenswrapper[4935]: E1005 07:50:42.507030 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="extract-utilities" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.507042 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="extract-utilities" Oct 05 07:50:42 crc kubenswrapper[4935]: E1005 07:50:42.507083 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="registry-server" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.507096 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="registry-server" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.507331 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8f18aa-57a1-4a86-bb30-f4b7511c31e1" containerName="registry-server" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.510134 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.525851 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jkpk"] Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.594906 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-utilities\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.594949 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmtgd\" (UniqueName: \"kubernetes.io/projected/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-kube-api-access-jmtgd\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.594999 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-catalog-content\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.696654 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-catalog-content\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.696791 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-utilities\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.696826 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmtgd\" (UniqueName: \"kubernetes.io/projected/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-kube-api-access-jmtgd\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.697246 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-catalog-content\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.697365 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-utilities\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.716335 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmtgd\" (UniqueName: \"kubernetes.io/projected/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-kube-api-access-jmtgd\") pod \"certified-operators-4jkpk\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:42 crc kubenswrapper[4935]: I1005 07:50:42.840965 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:43 crc kubenswrapper[4935]: I1005 07:50:43.344632 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jkpk"] Oct 05 07:50:43 crc kubenswrapper[4935]: I1005 07:50:43.646812 4935 generic.go:334] "Generic (PLEG): container finished" podID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerID="00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0" exitCode=0 Oct 05 07:50:43 crc kubenswrapper[4935]: I1005 07:50:43.647831 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jkpk" event={"ID":"02ff69de-4be1-42c5-ac7f-a07b67f35dfe","Type":"ContainerDied","Data":"00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0"} Oct 05 07:50:43 crc kubenswrapper[4935]: I1005 07:50:43.647939 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jkpk" event={"ID":"02ff69de-4be1-42c5-ac7f-a07b67f35dfe","Type":"ContainerStarted","Data":"e4f3dfaeb84893cc88951ea1befd4edb280a068811cc351bbfbb696fc317c877"} Oct 05 07:50:44 crc kubenswrapper[4935]: I1005 07:50:44.290194 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:50:44 crc kubenswrapper[4935]: I1005 07:50:44.290302 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:50:44 crc kubenswrapper[4935]: I1005 07:50:44.660386 4935 generic.go:334] "Generic (PLEG): container finished" podID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerID="ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160" exitCode=0 Oct 05 07:50:44 crc kubenswrapper[4935]: I1005 07:50:44.660444 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jkpk" event={"ID":"02ff69de-4be1-42c5-ac7f-a07b67f35dfe","Type":"ContainerDied","Data":"ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160"} Oct 05 07:50:45 crc kubenswrapper[4935]: I1005 07:50:45.671475 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jkpk" event={"ID":"02ff69de-4be1-42c5-ac7f-a07b67f35dfe","Type":"ContainerStarted","Data":"c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017"} Oct 05 07:50:45 crc kubenswrapper[4935]: I1005 07:50:45.692807 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4jkpk" podStartSLOduration=2.214732297 podStartE2EDuration="3.692786547s" podCreationTimestamp="2025-10-05 07:50:42 +0000 UTC" firstStartedPulling="2025-10-05 07:50:43.649698901 +0000 UTC m=+3477.532325361" lastFinishedPulling="2025-10-05 07:50:45.127753141 +0000 UTC m=+3479.010379611" observedRunningTime="2025-10-05 07:50:45.691467632 +0000 UTC m=+3479.574094112" watchObservedRunningTime="2025-10-05 07:50:45.692786547 +0000 UTC m=+3479.575412997" Oct 05 07:50:52 crc kubenswrapper[4935]: I1005 07:50:52.841884 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:52 crc kubenswrapper[4935]: I1005 07:50:52.842599 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:52 crc kubenswrapper[4935]: I1005 07:50:52.930209 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:53 crc kubenswrapper[4935]: I1005 07:50:53.796007 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:53 crc kubenswrapper[4935]: I1005 07:50:53.838188 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jkpk"] Oct 05 07:50:55 crc kubenswrapper[4935]: I1005 07:50:55.764569 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4jkpk" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerName="registry-server" containerID="cri-o://c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017" gracePeriod=2 Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.198420 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.300060 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmtgd\" (UniqueName: \"kubernetes.io/projected/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-kube-api-access-jmtgd\") pod \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.300146 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-catalog-content\") pod \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.300535 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-utilities\") pod \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\" (UID: \"02ff69de-4be1-42c5-ac7f-a07b67f35dfe\") " Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.302061 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-utilities" (OuterVolumeSpecName: "utilities") pod "02ff69de-4be1-42c5-ac7f-a07b67f35dfe" (UID: "02ff69de-4be1-42c5-ac7f-a07b67f35dfe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.316964 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-kube-api-access-jmtgd" (OuterVolumeSpecName: "kube-api-access-jmtgd") pod "02ff69de-4be1-42c5-ac7f-a07b67f35dfe" (UID: "02ff69de-4be1-42c5-ac7f-a07b67f35dfe"). InnerVolumeSpecName "kube-api-access-jmtgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.379884 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02ff69de-4be1-42c5-ac7f-a07b67f35dfe" (UID: "02ff69de-4be1-42c5-ac7f-a07b67f35dfe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.403100 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.403207 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmtgd\" (UniqueName: \"kubernetes.io/projected/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-kube-api-access-jmtgd\") on node \"crc\" DevicePath \"\"" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.403457 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ff69de-4be1-42c5-ac7f-a07b67f35dfe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.778612 4935 generic.go:334] "Generic (PLEG): container finished" podID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerID="c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017" exitCode=0 Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.778767 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jkpk" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.798478 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jkpk" event={"ID":"02ff69de-4be1-42c5-ac7f-a07b67f35dfe","Type":"ContainerDied","Data":"c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017"} Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.798540 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jkpk" event={"ID":"02ff69de-4be1-42c5-ac7f-a07b67f35dfe","Type":"ContainerDied","Data":"e4f3dfaeb84893cc88951ea1befd4edb280a068811cc351bbfbb696fc317c877"} Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.798572 4935 scope.go:117] "RemoveContainer" containerID="c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.836236 4935 scope.go:117] "RemoveContainer" containerID="ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.847001 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jkpk"] Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.856057 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4jkpk"] Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.877674 4935 scope.go:117] "RemoveContainer" containerID="00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.898878 4935 scope.go:117] "RemoveContainer" containerID="c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017" Oct 05 07:50:56 crc kubenswrapper[4935]: E1005 07:50:56.899804 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017\": container with ID starting with c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017 not found: ID does not exist" containerID="c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.899872 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017"} err="failed to get container status \"c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017\": rpc error: code = NotFound desc = could not find container \"c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017\": container with ID starting with c0a68c4a9dfb92ab534dd9d9e09841264fde8d6e195eb0f0505005e865a5a017 not found: ID does not exist" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.899952 4935 scope.go:117] "RemoveContainer" containerID="ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160" Oct 05 07:50:56 crc kubenswrapper[4935]: E1005 07:50:56.900448 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160\": container with ID starting with ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160 not found: ID does not exist" containerID="ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.900507 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160"} err="failed to get container status \"ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160\": rpc error: code = NotFound desc = could not find container \"ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160\": container with ID starting with ad31d463a6fd64827bc3918f62f1477016947613ef95f17b47ec8827dac31160 not found: ID does not exist" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.900543 4935 scope.go:117] "RemoveContainer" containerID="00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0" Oct 05 07:50:56 crc kubenswrapper[4935]: E1005 07:50:56.900961 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0\": container with ID starting with 00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0 not found: ID does not exist" containerID="00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0" Oct 05 07:50:56 crc kubenswrapper[4935]: I1005 07:50:56.901003 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0"} err="failed to get container status \"00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0\": rpc error: code = NotFound desc = could not find container \"00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0\": container with ID starting with 00c68dacdd264b8320065c67eb9883e5bfafacb8b4147e973f243c4faeda57e0 not found: ID does not exist" Oct 05 07:50:58 crc kubenswrapper[4935]: I1005 07:50:58.790403 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" path="/var/lib/kubelet/pods/02ff69de-4be1-42c5-ac7f-a07b67f35dfe/volumes" Oct 05 07:51:14 crc kubenswrapper[4935]: I1005 07:51:14.289117 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:51:14 crc kubenswrapper[4935]: I1005 07:51:14.289705 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:51:44 crc kubenswrapper[4935]: I1005 07:51:44.289176 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:51:44 crc kubenswrapper[4935]: I1005 07:51:44.289852 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:51:44 crc kubenswrapper[4935]: I1005 07:51:44.289943 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 07:51:44 crc kubenswrapper[4935]: I1005 07:51:44.290808 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:51:44 crc kubenswrapper[4935]: I1005 07:51:44.291040 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" gracePeriod=600 Oct 05 07:51:44 crc kubenswrapper[4935]: E1005 07:51:44.442266 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62823424_a6c9_4733_b50a_6f3ecb4a55a2.slice/crio-conmon-bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62823424_a6c9_4733_b50a_6f3ecb4a55a2.slice/crio-bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4.scope\": RecentStats: unable to find data in memory cache]" Oct 05 07:51:44 crc kubenswrapper[4935]: E1005 07:51:44.500765 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:51:45 crc kubenswrapper[4935]: I1005 07:51:45.235824 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" exitCode=0 Oct 05 07:51:45 crc kubenswrapper[4935]: I1005 07:51:45.235880 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4"} Oct 05 07:51:45 crc kubenswrapper[4935]: I1005 07:51:45.235984 4935 scope.go:117] "RemoveContainer" containerID="2142111d863206494d1c9e2a850cd11e2b5a731b9df5e3303b181065609022b4" Oct 05 07:51:45 crc kubenswrapper[4935]: I1005 07:51:45.236567 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:51:45 crc kubenswrapper[4935]: E1005 07:51:45.236986 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:51:57 crc kubenswrapper[4935]: I1005 07:51:57.777502 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:51:57 crc kubenswrapper[4935]: E1005 07:51:57.778149 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:52:12 crc kubenswrapper[4935]: I1005 07:52:12.777703 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:52:12 crc kubenswrapper[4935]: E1005 07:52:12.778819 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:52:24 crc kubenswrapper[4935]: I1005 07:52:24.777559 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:52:24 crc kubenswrapper[4935]: E1005 07:52:24.780413 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:52:37 crc kubenswrapper[4935]: I1005 07:52:37.778066 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:52:37 crc kubenswrapper[4935]: E1005 07:52:37.778738 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:52:51 crc kubenswrapper[4935]: I1005 07:52:51.777489 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:52:51 crc kubenswrapper[4935]: E1005 07:52:51.778492 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:53:06 crc kubenswrapper[4935]: I1005 07:53:06.785779 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:53:06 crc kubenswrapper[4935]: E1005 07:53:06.787154 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:53:21 crc kubenswrapper[4935]: I1005 07:53:21.778725 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:53:21 crc kubenswrapper[4935]: E1005 07:53:21.779480 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:53:35 crc kubenswrapper[4935]: I1005 07:53:35.777345 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:53:35 crc kubenswrapper[4935]: E1005 07:53:35.778389 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:53:46 crc kubenswrapper[4935]: I1005 07:53:46.784565 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:53:46 crc kubenswrapper[4935]: E1005 07:53:46.786455 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:54:00 crc kubenswrapper[4935]: I1005 07:54:00.777302 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:54:00 crc kubenswrapper[4935]: E1005 07:54:00.778278 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.028427 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ggsgw"] Oct 05 07:54:14 crc kubenswrapper[4935]: E1005 07:54:14.029667 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerName="extract-utilities" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.029687 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerName="extract-utilities" Oct 05 07:54:14 crc kubenswrapper[4935]: E1005 07:54:14.029742 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerName="registry-server" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.029752 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerName="registry-server" Oct 05 07:54:14 crc kubenswrapper[4935]: E1005 07:54:14.029766 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerName="extract-content" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.029774 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerName="extract-content" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.030168 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ff69de-4be1-42c5-ac7f-a07b67f35dfe" containerName="registry-server" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.031573 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.036445 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ggsgw"] Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.083422 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-utilities\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.083815 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-catalog-content\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.083998 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96ls7\" (UniqueName: \"kubernetes.io/projected/704fffda-337d-43bd-a79b-053aa17d3ee8-kube-api-access-96ls7\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.184886 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-catalog-content\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.184978 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96ls7\" (UniqueName: \"kubernetes.io/projected/704fffda-337d-43bd-a79b-053aa17d3ee8-kube-api-access-96ls7\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.185040 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-utilities\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.185770 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-utilities\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.185764 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-catalog-content\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.210048 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96ls7\" (UniqueName: \"kubernetes.io/projected/704fffda-337d-43bd-a79b-053aa17d3ee8-kube-api-access-96ls7\") pod \"redhat-marketplace-ggsgw\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.370679 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:14 crc kubenswrapper[4935]: I1005 07:54:14.790381 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ggsgw"] Oct 05 07:54:15 crc kubenswrapper[4935]: I1005 07:54:15.596333 4935 generic.go:334] "Generic (PLEG): container finished" podID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerID="5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1" exitCode=0 Oct 05 07:54:15 crc kubenswrapper[4935]: I1005 07:54:15.596432 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ggsgw" event={"ID":"704fffda-337d-43bd-a79b-053aa17d3ee8","Type":"ContainerDied","Data":"5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1"} Oct 05 07:54:15 crc kubenswrapper[4935]: I1005 07:54:15.596635 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ggsgw" event={"ID":"704fffda-337d-43bd-a79b-053aa17d3ee8","Type":"ContainerStarted","Data":"9e808819adb292089a37dcf6cdaf130368080ef7194259dbbe75576bc8e8e27c"} Oct 05 07:54:15 crc kubenswrapper[4935]: I1005 07:54:15.599582 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:54:15 crc kubenswrapper[4935]: I1005 07:54:15.777209 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:54:15 crc kubenswrapper[4935]: E1005 07:54:15.777753 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:54:16 crc kubenswrapper[4935]: I1005 07:54:16.606084 4935 generic.go:334] "Generic (PLEG): container finished" podID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerID="c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d" exitCode=0 Oct 05 07:54:16 crc kubenswrapper[4935]: I1005 07:54:16.606475 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ggsgw" event={"ID":"704fffda-337d-43bd-a79b-053aa17d3ee8","Type":"ContainerDied","Data":"c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d"} Oct 05 07:54:17 crc kubenswrapper[4935]: I1005 07:54:17.622162 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ggsgw" event={"ID":"704fffda-337d-43bd-a79b-053aa17d3ee8","Type":"ContainerStarted","Data":"6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429"} Oct 05 07:54:17 crc kubenswrapper[4935]: I1005 07:54:17.650561 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ggsgw" podStartSLOduration=3.267975707 podStartE2EDuration="4.650541463s" podCreationTimestamp="2025-10-05 07:54:13 +0000 UTC" firstStartedPulling="2025-10-05 07:54:15.599368442 +0000 UTC m=+3689.481994902" lastFinishedPulling="2025-10-05 07:54:16.981934198 +0000 UTC m=+3690.864560658" observedRunningTime="2025-10-05 07:54:17.646732302 +0000 UTC m=+3691.529358792" watchObservedRunningTime="2025-10-05 07:54:17.650541463 +0000 UTC m=+3691.533167913" Oct 05 07:54:24 crc kubenswrapper[4935]: I1005 07:54:24.371153 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:24 crc kubenswrapper[4935]: I1005 07:54:24.371741 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:24 crc kubenswrapper[4935]: I1005 07:54:24.461458 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:24 crc kubenswrapper[4935]: I1005 07:54:24.757330 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:24 crc kubenswrapper[4935]: I1005 07:54:24.811859 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ggsgw"] Oct 05 07:54:26 crc kubenswrapper[4935]: I1005 07:54:26.696335 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ggsgw" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerName="registry-server" containerID="cri-o://6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429" gracePeriod=2 Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.092026 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.178710 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-utilities\") pod \"704fffda-337d-43bd-a79b-053aa17d3ee8\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.178775 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-catalog-content\") pod \"704fffda-337d-43bd-a79b-053aa17d3ee8\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.178836 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96ls7\" (UniqueName: \"kubernetes.io/projected/704fffda-337d-43bd-a79b-053aa17d3ee8-kube-api-access-96ls7\") pod \"704fffda-337d-43bd-a79b-053aa17d3ee8\" (UID: \"704fffda-337d-43bd-a79b-053aa17d3ee8\") " Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.179414 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-utilities" (OuterVolumeSpecName: "utilities") pod "704fffda-337d-43bd-a79b-053aa17d3ee8" (UID: "704fffda-337d-43bd-a79b-053aa17d3ee8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.191168 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/704fffda-337d-43bd-a79b-053aa17d3ee8-kube-api-access-96ls7" (OuterVolumeSpecName: "kube-api-access-96ls7") pod "704fffda-337d-43bd-a79b-053aa17d3ee8" (UID: "704fffda-337d-43bd-a79b-053aa17d3ee8"). InnerVolumeSpecName "kube-api-access-96ls7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.196392 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "704fffda-337d-43bd-a79b-053aa17d3ee8" (UID: "704fffda-337d-43bd-a79b-053aa17d3ee8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.279990 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96ls7\" (UniqueName: \"kubernetes.io/projected/704fffda-337d-43bd-a79b-053aa17d3ee8-kube-api-access-96ls7\") on node \"crc\" DevicePath \"\"" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.280513 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.280602 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/704fffda-337d-43bd-a79b-053aa17d3ee8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.708801 4935 generic.go:334] "Generic (PLEG): container finished" podID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerID="6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429" exitCode=0 Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.708860 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ggsgw" event={"ID":"704fffda-337d-43bd-a79b-053aa17d3ee8","Type":"ContainerDied","Data":"6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429"} Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.708916 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ggsgw" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.708935 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ggsgw" event={"ID":"704fffda-337d-43bd-a79b-053aa17d3ee8","Type":"ContainerDied","Data":"9e808819adb292089a37dcf6cdaf130368080ef7194259dbbe75576bc8e8e27c"} Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.708960 4935 scope.go:117] "RemoveContainer" containerID="6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.744830 4935 scope.go:117] "RemoveContainer" containerID="c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.759093 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ggsgw"] Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.776499 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ggsgw"] Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.788526 4935 scope.go:117] "RemoveContainer" containerID="5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.813950 4935 scope.go:117] "RemoveContainer" containerID="6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429" Oct 05 07:54:27 crc kubenswrapper[4935]: E1005 07:54:27.814719 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429\": container with ID starting with 6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429 not found: ID does not exist" containerID="6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.814834 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429"} err="failed to get container status \"6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429\": rpc error: code = NotFound desc = could not find container \"6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429\": container with ID starting with 6ef5756605437227e6923955547ccf5b8f5982c6b6ff4b4e0995dadda6f4a429 not found: ID does not exist" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.814988 4935 scope.go:117] "RemoveContainer" containerID="c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d" Oct 05 07:54:27 crc kubenswrapper[4935]: E1005 07:54:27.815403 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d\": container with ID starting with c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d not found: ID does not exist" containerID="c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.815459 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d"} err="failed to get container status \"c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d\": rpc error: code = NotFound desc = could not find container \"c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d\": container with ID starting with c3e55f86ecb01a9920306136120e6631d00efc77b6103031132d2bbe71ddf21d not found: ID does not exist" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.815493 4935 scope.go:117] "RemoveContainer" containerID="5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1" Oct 05 07:54:27 crc kubenswrapper[4935]: E1005 07:54:27.815835 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1\": container with ID starting with 5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1 not found: ID does not exist" containerID="5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1" Oct 05 07:54:27 crc kubenswrapper[4935]: I1005 07:54:27.815871 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1"} err="failed to get container status \"5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1\": rpc error: code = NotFound desc = could not find container \"5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1\": container with ID starting with 5db989874ccbb6560b66b9cbb022d2129fe60a9f2cb57620c251f4acca541aa1 not found: ID does not exist" Oct 05 07:54:27 crc kubenswrapper[4935]: E1005 07:54:27.858622 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod704fffda_337d_43bd_a79b_053aa17d3ee8.slice/crio-9e808819adb292089a37dcf6cdaf130368080ef7194259dbbe75576bc8e8e27c\": RecentStats: unable to find data in memory cache]" Oct 05 07:54:28 crc kubenswrapper[4935]: I1005 07:54:28.788863 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" path="/var/lib/kubelet/pods/704fffda-337d-43bd-a79b-053aa17d3ee8/volumes" Oct 05 07:54:29 crc kubenswrapper[4935]: I1005 07:54:29.777741 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:54:29 crc kubenswrapper[4935]: E1005 07:54:29.778485 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:54:42 crc kubenswrapper[4935]: I1005 07:54:42.777160 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:54:42 crc kubenswrapper[4935]: E1005 07:54:42.778019 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:54:55 crc kubenswrapper[4935]: I1005 07:54:55.777666 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:54:55 crc kubenswrapper[4935]: E1005 07:54:55.778524 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:55:10 crc kubenswrapper[4935]: I1005 07:55:10.777469 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:55:10 crc kubenswrapper[4935]: E1005 07:55:10.779447 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:55:25 crc kubenswrapper[4935]: I1005 07:55:25.778002 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:55:25 crc kubenswrapper[4935]: E1005 07:55:25.780411 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:55:37 crc kubenswrapper[4935]: I1005 07:55:37.777411 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:55:37 crc kubenswrapper[4935]: E1005 07:55:37.778652 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:55:51 crc kubenswrapper[4935]: I1005 07:55:51.777613 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:55:51 crc kubenswrapper[4935]: E1005 07:55:51.778678 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:56:05 crc kubenswrapper[4935]: I1005 07:56:05.777175 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:56:05 crc kubenswrapper[4935]: E1005 07:56:05.778304 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:56:17 crc kubenswrapper[4935]: I1005 07:56:17.777214 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:56:17 crc kubenswrapper[4935]: E1005 07:56:17.778104 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:56:31 crc kubenswrapper[4935]: I1005 07:56:31.777569 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:56:31 crc kubenswrapper[4935]: E1005 07:56:31.778813 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 07:56:45 crc kubenswrapper[4935]: I1005 07:56:45.777410 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 07:56:46 crc kubenswrapper[4935]: I1005 07:56:46.045196 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"01e85d791e8730d0b4166d9cd3de9cf5c5a5f22d1c7dfc35c0721f631ba9149d"} Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.319879 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gh9ks"] Oct 05 07:57:23 crc kubenswrapper[4935]: E1005 07:57:23.321012 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerName="registry-server" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.321035 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerName="registry-server" Oct 05 07:57:23 crc kubenswrapper[4935]: E1005 07:57:23.321090 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerName="extract-content" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.321103 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerName="extract-content" Oct 05 07:57:23 crc kubenswrapper[4935]: E1005 07:57:23.321122 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerName="extract-utilities" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.321135 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerName="extract-utilities" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.321400 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="704fffda-337d-43bd-a79b-053aa17d3ee8" containerName="registry-server" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.323349 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.335849 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gh9ks"] Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.361434 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97wbz\" (UniqueName: \"kubernetes.io/projected/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-kube-api-access-97wbz\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.361515 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-catalog-content\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.361545 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-utilities\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.462857 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97wbz\" (UniqueName: \"kubernetes.io/projected/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-kube-api-access-97wbz\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.462971 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-catalog-content\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.463000 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-utilities\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.463645 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-utilities\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.463778 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-catalog-content\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.492072 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97wbz\" (UniqueName: \"kubernetes.io/projected/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-kube-api-access-97wbz\") pod \"community-operators-gh9ks\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:23 crc kubenswrapper[4935]: I1005 07:57:23.649616 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:24 crc kubenswrapper[4935]: I1005 07:57:24.122630 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gh9ks"] Oct 05 07:57:24 crc kubenswrapper[4935]: I1005 07:57:24.373421 4935 generic.go:334] "Generic (PLEG): container finished" podID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerID="d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326" exitCode=0 Oct 05 07:57:24 crc kubenswrapper[4935]: I1005 07:57:24.373478 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9ks" event={"ID":"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272","Type":"ContainerDied","Data":"d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326"} Oct 05 07:57:24 crc kubenswrapper[4935]: I1005 07:57:24.373699 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9ks" event={"ID":"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272","Type":"ContainerStarted","Data":"140b514b10288835a082ba6f3d3e3e78ab2f511434bd180463ad83f3f9ac23b5"} Oct 05 07:57:25 crc kubenswrapper[4935]: I1005 07:57:25.384254 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9ks" event={"ID":"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272","Type":"ContainerStarted","Data":"9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2"} Oct 05 07:57:26 crc kubenswrapper[4935]: I1005 07:57:26.393853 4935 generic.go:334] "Generic (PLEG): container finished" podID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerID="9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2" exitCode=0 Oct 05 07:57:26 crc kubenswrapper[4935]: I1005 07:57:26.393938 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9ks" event={"ID":"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272","Type":"ContainerDied","Data":"9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2"} Oct 05 07:57:27 crc kubenswrapper[4935]: I1005 07:57:27.411079 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9ks" event={"ID":"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272","Type":"ContainerStarted","Data":"b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b"} Oct 05 07:57:27 crc kubenswrapper[4935]: I1005 07:57:27.456836 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gh9ks" podStartSLOduration=2.010083256 podStartE2EDuration="4.45678552s" podCreationTimestamp="2025-10-05 07:57:23 +0000 UTC" firstStartedPulling="2025-10-05 07:57:24.375282491 +0000 UTC m=+3878.257908961" lastFinishedPulling="2025-10-05 07:57:26.821984755 +0000 UTC m=+3880.704611225" observedRunningTime="2025-10-05 07:57:27.43424146 +0000 UTC m=+3881.316867950" watchObservedRunningTime="2025-10-05 07:57:27.45678552 +0000 UTC m=+3881.339412020" Oct 05 07:57:33 crc kubenswrapper[4935]: I1005 07:57:33.650948 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:33 crc kubenswrapper[4935]: I1005 07:57:33.653195 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:33 crc kubenswrapper[4935]: I1005 07:57:33.724368 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:34 crc kubenswrapper[4935]: I1005 07:57:34.543156 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:34 crc kubenswrapper[4935]: I1005 07:57:34.616106 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gh9ks"] Oct 05 07:57:36 crc kubenswrapper[4935]: I1005 07:57:36.487948 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gh9ks" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerName="registry-server" containerID="cri-o://b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b" gracePeriod=2 Oct 05 07:57:36 crc kubenswrapper[4935]: I1005 07:57:36.914225 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.000544 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-catalog-content\") pod \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.000740 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97wbz\" (UniqueName: \"kubernetes.io/projected/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-kube-api-access-97wbz\") pod \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.000827 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-utilities\") pod \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\" (UID: \"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272\") " Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.003208 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-utilities" (OuterVolumeSpecName: "utilities") pod "1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" (UID: "1f9dc4d4-7d32-4d2e-af5a-c71e175e9272"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.011459 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-kube-api-access-97wbz" (OuterVolumeSpecName: "kube-api-access-97wbz") pod "1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" (UID: "1f9dc4d4-7d32-4d2e-af5a-c71e175e9272"). InnerVolumeSpecName "kube-api-access-97wbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.067548 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" (UID: "1f9dc4d4-7d32-4d2e-af5a-c71e175e9272"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.103464 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.103504 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.103520 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97wbz\" (UniqueName: \"kubernetes.io/projected/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272-kube-api-access-97wbz\") on node \"crc\" DevicePath \"\"" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.498698 4935 generic.go:334] "Generic (PLEG): container finished" podID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerID="b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b" exitCode=0 Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.498762 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gh9ks" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.498760 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9ks" event={"ID":"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272","Type":"ContainerDied","Data":"b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b"} Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.500131 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9ks" event={"ID":"1f9dc4d4-7d32-4d2e-af5a-c71e175e9272","Type":"ContainerDied","Data":"140b514b10288835a082ba6f3d3e3e78ab2f511434bd180463ad83f3f9ac23b5"} Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.500179 4935 scope.go:117] "RemoveContainer" containerID="b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.538207 4935 scope.go:117] "RemoveContainer" containerID="9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.566209 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gh9ks"] Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.578186 4935 scope.go:117] "RemoveContainer" containerID="d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.582318 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gh9ks"] Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.607725 4935 scope.go:117] "RemoveContainer" containerID="b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b" Oct 05 07:57:37 crc kubenswrapper[4935]: E1005 07:57:37.608511 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b\": container with ID starting with b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b not found: ID does not exist" containerID="b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.608564 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b"} err="failed to get container status \"b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b\": rpc error: code = NotFound desc = could not find container \"b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b\": container with ID starting with b3aa7a1a62630c5e740b51a6f1c571b8cc95d19dea9c830e4da32178c829572b not found: ID does not exist" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.608596 4935 scope.go:117] "RemoveContainer" containerID="9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2" Oct 05 07:57:37 crc kubenswrapper[4935]: E1005 07:57:37.609112 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2\": container with ID starting with 9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2 not found: ID does not exist" containerID="9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.609144 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2"} err="failed to get container status \"9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2\": rpc error: code = NotFound desc = could not find container \"9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2\": container with ID starting with 9ff9fc80a424d72cf1ec1dc885c35be85c04f2b839fded2196edce0d19c193b2 not found: ID does not exist" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.609162 4935 scope.go:117] "RemoveContainer" containerID="d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326" Oct 05 07:57:37 crc kubenswrapper[4935]: E1005 07:57:37.609516 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326\": container with ID starting with d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326 not found: ID does not exist" containerID="d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326" Oct 05 07:57:37 crc kubenswrapper[4935]: I1005 07:57:37.609666 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326"} err="failed to get container status \"d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326\": rpc error: code = NotFound desc = could not find container \"d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326\": container with ID starting with d5977d2d928924245b00fa8d8e4b2b6d49371752bcb13fdff95a3752d9d3f326 not found: ID does not exist" Oct 05 07:57:38 crc kubenswrapper[4935]: I1005 07:57:38.794484 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" path="/var/lib/kubelet/pods/1f9dc4d4-7d32-4d2e-af5a-c71e175e9272/volumes" Oct 05 07:59:14 crc kubenswrapper[4935]: I1005 07:59:14.289406 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:59:14 crc kubenswrapper[4935]: I1005 07:59:14.289928 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:59:44 crc kubenswrapper[4935]: I1005 07:59:44.289750 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:59:44 crc kubenswrapper[4935]: I1005 07:59:44.290311 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.168858 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b"] Oct 05 08:00:00 crc kubenswrapper[4935]: E1005 08:00:00.169636 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerName="extract-utilities" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.169650 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerName="extract-utilities" Oct 05 08:00:00 crc kubenswrapper[4935]: E1005 08:00:00.169677 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerName="extract-content" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.169685 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerName="extract-content" Oct 05 08:00:00 crc kubenswrapper[4935]: E1005 08:00:00.169694 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerName="registry-server" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.169699 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerName="registry-server" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.169844 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9dc4d4-7d32-4d2e-af5a-c71e175e9272" containerName="registry-server" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.170440 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.175317 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.175537 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.183430 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b"] Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.290541 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c163e86-9707-4230-8144-0016dd2b6c74-secret-volume\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.290662 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klmp5\" (UniqueName: \"kubernetes.io/projected/7c163e86-9707-4230-8144-0016dd2b6c74-kube-api-access-klmp5\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.290767 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c163e86-9707-4230-8144-0016dd2b6c74-config-volume\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.372156 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s4gmv"] Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.373968 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.392114 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-catalog-content\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.392168 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvwdj\" (UniqueName: \"kubernetes.io/projected/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-kube-api-access-rvwdj\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.392232 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klmp5\" (UniqueName: \"kubernetes.io/projected/7c163e86-9707-4230-8144-0016dd2b6c74-kube-api-access-klmp5\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.392282 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c163e86-9707-4230-8144-0016dd2b6c74-config-volume\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.392325 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-utilities\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.392362 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c163e86-9707-4230-8144-0016dd2b6c74-secret-volume\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.393424 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c163e86-9707-4230-8144-0016dd2b6c74-config-volume\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.393463 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s4gmv"] Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.402429 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c163e86-9707-4230-8144-0016dd2b6c74-secret-volume\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.415276 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klmp5\" (UniqueName: \"kubernetes.io/projected/7c163e86-9707-4230-8144-0016dd2b6c74-kube-api-access-klmp5\") pod \"collect-profiles-29327520-s4x7b\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.493250 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvwdj\" (UniqueName: \"kubernetes.io/projected/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-kube-api-access-rvwdj\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.493398 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-utilities\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.493438 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-catalog-content\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.494078 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-catalog-content\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.494142 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-utilities\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.497660 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.521329 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvwdj\" (UniqueName: \"kubernetes.io/projected/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-kube-api-access-rvwdj\") pod \"redhat-operators-s4gmv\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.694309 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:00 crc kubenswrapper[4935]: I1005 08:00:00.918911 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b"] Oct 05 08:00:01 crc kubenswrapper[4935]: I1005 08:00:01.116308 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s4gmv"] Oct 05 08:00:01 crc kubenswrapper[4935]: W1005 08:00:01.123411 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7de9c98f_d6f9_4f8b_b8f5_395b0445d59b.slice/crio-41a4f1bad47e5d7280f37e5ae2b848e151550c9787126c23281b96930e3b1a2a WatchSource:0}: Error finding container 41a4f1bad47e5d7280f37e5ae2b848e151550c9787126c23281b96930e3b1a2a: Status 404 returned error can't find the container with id 41a4f1bad47e5d7280f37e5ae2b848e151550c9787126c23281b96930e3b1a2a Oct 05 08:00:01 crc kubenswrapper[4935]: I1005 08:00:01.803618 4935 generic.go:334] "Generic (PLEG): container finished" podID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerID="8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c" exitCode=0 Oct 05 08:00:01 crc kubenswrapper[4935]: I1005 08:00:01.804159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4gmv" event={"ID":"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b","Type":"ContainerDied","Data":"8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c"} Oct 05 08:00:01 crc kubenswrapper[4935]: I1005 08:00:01.804239 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4gmv" event={"ID":"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b","Type":"ContainerStarted","Data":"41a4f1bad47e5d7280f37e5ae2b848e151550c9787126c23281b96930e3b1a2a"} Oct 05 08:00:01 crc kubenswrapper[4935]: I1005 08:00:01.806404 4935 generic.go:334] "Generic (PLEG): container finished" podID="7c163e86-9707-4230-8144-0016dd2b6c74" containerID="5ebe97d01e4970397aeb73ed4255c3c70f57a88ade2bb6699d24142822b5645b" exitCode=0 Oct 05 08:00:01 crc kubenswrapper[4935]: I1005 08:00:01.806455 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" event={"ID":"7c163e86-9707-4230-8144-0016dd2b6c74","Type":"ContainerDied","Data":"5ebe97d01e4970397aeb73ed4255c3c70f57a88ade2bb6699d24142822b5645b"} Oct 05 08:00:01 crc kubenswrapper[4935]: I1005 08:00:01.806490 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" event={"ID":"7c163e86-9707-4230-8144-0016dd2b6c74","Type":"ContainerStarted","Data":"4815a1993138cdeb9981f9bcc3a60f9c29cdac9eac716803d2da1fc9f290723f"} Oct 05 08:00:01 crc kubenswrapper[4935]: I1005 08:00:01.806860 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.089069 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.131652 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klmp5\" (UniqueName: \"kubernetes.io/projected/7c163e86-9707-4230-8144-0016dd2b6c74-kube-api-access-klmp5\") pod \"7c163e86-9707-4230-8144-0016dd2b6c74\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.131926 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c163e86-9707-4230-8144-0016dd2b6c74-config-volume\") pod \"7c163e86-9707-4230-8144-0016dd2b6c74\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.131981 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c163e86-9707-4230-8144-0016dd2b6c74-secret-volume\") pod \"7c163e86-9707-4230-8144-0016dd2b6c74\" (UID: \"7c163e86-9707-4230-8144-0016dd2b6c74\") " Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.132571 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c163e86-9707-4230-8144-0016dd2b6c74-config-volume" (OuterVolumeSpecName: "config-volume") pod "7c163e86-9707-4230-8144-0016dd2b6c74" (UID: "7c163e86-9707-4230-8144-0016dd2b6c74"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.160750 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c163e86-9707-4230-8144-0016dd2b6c74-kube-api-access-klmp5" (OuterVolumeSpecName: "kube-api-access-klmp5") pod "7c163e86-9707-4230-8144-0016dd2b6c74" (UID: "7c163e86-9707-4230-8144-0016dd2b6c74"). InnerVolumeSpecName "kube-api-access-klmp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.161053 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c163e86-9707-4230-8144-0016dd2b6c74-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7c163e86-9707-4230-8144-0016dd2b6c74" (UID: "7c163e86-9707-4230-8144-0016dd2b6c74"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.233157 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c163e86-9707-4230-8144-0016dd2b6c74-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.233191 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c163e86-9707-4230-8144-0016dd2b6c74-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.233203 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klmp5\" (UniqueName: \"kubernetes.io/projected/7c163e86-9707-4230-8144-0016dd2b6c74-kube-api-access-klmp5\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.833195 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" event={"ID":"7c163e86-9707-4230-8144-0016dd2b6c74","Type":"ContainerDied","Data":"4815a1993138cdeb9981f9bcc3a60f9c29cdac9eac716803d2da1fc9f290723f"} Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.833246 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b" Oct 05 08:00:03 crc kubenswrapper[4935]: I1005 08:00:03.833262 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4815a1993138cdeb9981f9bcc3a60f9c29cdac9eac716803d2da1fc9f290723f" Oct 05 08:00:04 crc kubenswrapper[4935]: I1005 08:00:04.153250 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz"] Oct 05 08:00:04 crc kubenswrapper[4935]: I1005 08:00:04.157992 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-gkbbz"] Oct 05 08:00:04 crc kubenswrapper[4935]: I1005 08:00:04.787885 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a" path="/var/lib/kubelet/pods/20f3b4e0-3e03-4d21-8d6c-8a2babf28f1a/volumes" Oct 05 08:00:04 crc kubenswrapper[4935]: I1005 08:00:04.840855 4935 generic.go:334] "Generic (PLEG): container finished" podID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerID="fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c" exitCode=0 Oct 05 08:00:04 crc kubenswrapper[4935]: I1005 08:00:04.840998 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4gmv" event={"ID":"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b","Type":"ContainerDied","Data":"fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c"} Oct 05 08:00:05 crc kubenswrapper[4935]: I1005 08:00:05.851204 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4gmv" event={"ID":"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b","Type":"ContainerStarted","Data":"a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035"} Oct 05 08:00:05 crc kubenswrapper[4935]: I1005 08:00:05.871740 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s4gmv" podStartSLOduration=2.3816891780000002 podStartE2EDuration="5.871723808s" podCreationTimestamp="2025-10-05 08:00:00 +0000 UTC" firstStartedPulling="2025-10-05 08:00:01.806633835 +0000 UTC m=+4035.689260295" lastFinishedPulling="2025-10-05 08:00:05.296668455 +0000 UTC m=+4039.179294925" observedRunningTime="2025-10-05 08:00:05.868149023 +0000 UTC m=+4039.750775483" watchObservedRunningTime="2025-10-05 08:00:05.871723808 +0000 UTC m=+4039.754350268" Oct 05 08:00:10 crc kubenswrapper[4935]: I1005 08:00:10.695677 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:10 crc kubenswrapper[4935]: I1005 08:00:10.696242 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:10 crc kubenswrapper[4935]: I1005 08:00:10.794867 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:10 crc kubenswrapper[4935]: I1005 08:00:10.929422 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:11 crc kubenswrapper[4935]: I1005 08:00:11.024982 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s4gmv"] Oct 05 08:00:12 crc kubenswrapper[4935]: I1005 08:00:12.900261 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s4gmv" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerName="registry-server" containerID="cri-o://a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035" gracePeriod=2 Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.430024 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.585274 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-utilities\") pod \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.585327 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-catalog-content\") pod \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.585388 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvwdj\" (UniqueName: \"kubernetes.io/projected/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-kube-api-access-rvwdj\") pod \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\" (UID: \"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b\") " Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.586063 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-utilities" (OuterVolumeSpecName: "utilities") pod "7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" (UID: "7de9c98f-d6f9-4f8b-b8f5-395b0445d59b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.590154 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-kube-api-access-rvwdj" (OuterVolumeSpecName: "kube-api-access-rvwdj") pod "7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" (UID: "7de9c98f-d6f9-4f8b-b8f5-395b0445d59b"). InnerVolumeSpecName "kube-api-access-rvwdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.687166 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvwdj\" (UniqueName: \"kubernetes.io/projected/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-kube-api-access-rvwdj\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.687238 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.691560 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" (UID: "7de9c98f-d6f9-4f8b-b8f5-395b0445d59b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.788991 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.909115 4935 generic.go:334] "Generic (PLEG): container finished" podID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerID="a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035" exitCode=0 Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.909198 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4gmv" event={"ID":"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b","Type":"ContainerDied","Data":"a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035"} Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.909429 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4gmv" event={"ID":"7de9c98f-d6f9-4f8b-b8f5-395b0445d59b","Type":"ContainerDied","Data":"41a4f1bad47e5d7280f37e5ae2b848e151550c9787126c23281b96930e3b1a2a"} Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.909447 4935 scope.go:117] "RemoveContainer" containerID="a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.909202 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4gmv" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.933545 4935 scope.go:117] "RemoveContainer" containerID="fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.944990 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s4gmv"] Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.952516 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s4gmv"] Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.961837 4935 scope.go:117] "RemoveContainer" containerID="8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.989252 4935 scope.go:117] "RemoveContainer" containerID="a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035" Oct 05 08:00:13 crc kubenswrapper[4935]: E1005 08:00:13.989622 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035\": container with ID starting with a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035 not found: ID does not exist" containerID="a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.989665 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035"} err="failed to get container status \"a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035\": rpc error: code = NotFound desc = could not find container \"a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035\": container with ID starting with a015a42ae73117089865a6532d6bf4a7e6a2e581c6f92dfefe07f9307b56f035 not found: ID does not exist" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.989692 4935 scope.go:117] "RemoveContainer" containerID="fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c" Oct 05 08:00:13 crc kubenswrapper[4935]: E1005 08:00:13.990075 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c\": container with ID starting with fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c not found: ID does not exist" containerID="fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.990121 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c"} err="failed to get container status \"fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c\": rpc error: code = NotFound desc = could not find container \"fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c\": container with ID starting with fdf6354fb415e6324eb47b579aec36d9a92532dcc139c67edeee7ec2b805b81c not found: ID does not exist" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.990150 4935 scope.go:117] "RemoveContainer" containerID="8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c" Oct 05 08:00:13 crc kubenswrapper[4935]: E1005 08:00:13.990643 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c\": container with ID starting with 8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c not found: ID does not exist" containerID="8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c" Oct 05 08:00:13 crc kubenswrapper[4935]: I1005 08:00:13.990667 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c"} err="failed to get container status \"8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c\": rpc error: code = NotFound desc = could not find container \"8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c\": container with ID starting with 8fad3e21ffbe6511fada091fdd200943dbec5193e01a25ef4947e4d45c224d3c not found: ID does not exist" Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.289669 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.289724 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.289762 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.290372 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01e85d791e8730d0b4166d9cd3de9cf5c5a5f22d1c7dfc35c0721f631ba9149d"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.290440 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://01e85d791e8730d0b4166d9cd3de9cf5c5a5f22d1c7dfc35c0721f631ba9149d" gracePeriod=600 Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.787081 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" path="/var/lib/kubelet/pods/7de9c98f-d6f9-4f8b-b8f5-395b0445d59b/volumes" Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.917638 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="01e85d791e8730d0b4166d9cd3de9cf5c5a5f22d1c7dfc35c0721f631ba9149d" exitCode=0 Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.917710 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"01e85d791e8730d0b4166d9cd3de9cf5c5a5f22d1c7dfc35c0721f631ba9149d"} Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.917825 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c"} Oct 05 08:00:14 crc kubenswrapper[4935]: I1005 08:00:14.917850 4935 scope.go:117] "RemoveContainer" containerID="bc9dee056b7f5184cdc01ba2f53b96d05ae20291de55f47c925faad2e0e67cb4" Oct 05 08:00:55 crc kubenswrapper[4935]: I1005 08:00:55.522601 4935 scope.go:117] "RemoveContainer" containerID="e5dc00576705b6f955a15b49132057c1684260abcf8020c2a1cb6e30356b9f6a" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.972176 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fljbf"] Oct 05 08:02:07 crc kubenswrapper[4935]: E1005 08:02:07.973234 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerName="extract-content" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.973256 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerName="extract-content" Oct 05 08:02:07 crc kubenswrapper[4935]: E1005 08:02:07.973299 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerName="extract-utilities" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.973310 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerName="extract-utilities" Oct 05 08:02:07 crc kubenswrapper[4935]: E1005 08:02:07.973323 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerName="registry-server" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.973336 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerName="registry-server" Oct 05 08:02:07 crc kubenswrapper[4935]: E1005 08:02:07.973364 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c163e86-9707-4230-8144-0016dd2b6c74" containerName="collect-profiles" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.973374 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c163e86-9707-4230-8144-0016dd2b6c74" containerName="collect-profiles" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.973569 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c163e86-9707-4230-8144-0016dd2b6c74" containerName="collect-profiles" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.973605 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7de9c98f-d6f9-4f8b-b8f5-395b0445d59b" containerName="registry-server" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.975135 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:07 crc kubenswrapper[4935]: I1005 08:02:07.987762 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fljbf"] Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.025258 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfph5\" (UniqueName: \"kubernetes.io/projected/591a2fb1-8a31-4a87-8f5d-26d111519c78-kube-api-access-cfph5\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.025326 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-utilities\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.025369 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-catalog-content\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.126504 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfph5\" (UniqueName: \"kubernetes.io/projected/591a2fb1-8a31-4a87-8f5d-26d111519c78-kube-api-access-cfph5\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.126567 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-utilities\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.126607 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-catalog-content\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.127214 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-utilities\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.127229 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-catalog-content\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.160986 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfph5\" (UniqueName: \"kubernetes.io/projected/591a2fb1-8a31-4a87-8f5d-26d111519c78-kube-api-access-cfph5\") pod \"certified-operators-fljbf\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.310199 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.827381 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fljbf"] Oct 05 08:02:08 crc kubenswrapper[4935]: I1005 08:02:08.927418 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fljbf" event={"ID":"591a2fb1-8a31-4a87-8f5d-26d111519c78","Type":"ContainerStarted","Data":"470a5f7315536dfbc7a5ccc06a106c24da549a5ae95f3c97af6b1409d56b1f1a"} Oct 05 08:02:09 crc kubenswrapper[4935]: I1005 08:02:09.936715 4935 generic.go:334] "Generic (PLEG): container finished" podID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerID="9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91" exitCode=0 Oct 05 08:02:09 crc kubenswrapper[4935]: I1005 08:02:09.936767 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fljbf" event={"ID":"591a2fb1-8a31-4a87-8f5d-26d111519c78","Type":"ContainerDied","Data":"9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91"} Oct 05 08:02:10 crc kubenswrapper[4935]: I1005 08:02:10.947988 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fljbf" event={"ID":"591a2fb1-8a31-4a87-8f5d-26d111519c78","Type":"ContainerStarted","Data":"6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8"} Oct 05 08:02:11 crc kubenswrapper[4935]: I1005 08:02:11.956026 4935 generic.go:334] "Generic (PLEG): container finished" podID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerID="6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8" exitCode=0 Oct 05 08:02:11 crc kubenswrapper[4935]: I1005 08:02:11.956075 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fljbf" event={"ID":"591a2fb1-8a31-4a87-8f5d-26d111519c78","Type":"ContainerDied","Data":"6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8"} Oct 05 08:02:12 crc kubenswrapper[4935]: I1005 08:02:12.967264 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fljbf" event={"ID":"591a2fb1-8a31-4a87-8f5d-26d111519c78","Type":"ContainerStarted","Data":"f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804"} Oct 05 08:02:13 crc kubenswrapper[4935]: I1005 08:02:12.999870 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fljbf" podStartSLOduration=3.549048407 podStartE2EDuration="5.999838182s" podCreationTimestamp="2025-10-05 08:02:07 +0000 UTC" firstStartedPulling="2025-10-05 08:02:09.939258439 +0000 UTC m=+4163.821884919" lastFinishedPulling="2025-10-05 08:02:12.390048234 +0000 UTC m=+4166.272674694" observedRunningTime="2025-10-05 08:02:12.994012766 +0000 UTC m=+4166.876639246" watchObservedRunningTime="2025-10-05 08:02:12.999838182 +0000 UTC m=+4166.882464682" Oct 05 08:02:14 crc kubenswrapper[4935]: I1005 08:02:14.290042 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:02:14 crc kubenswrapper[4935]: I1005 08:02:14.290134 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:02:18 crc kubenswrapper[4935]: I1005 08:02:18.310382 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:18 crc kubenswrapper[4935]: I1005 08:02:18.311646 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:18 crc kubenswrapper[4935]: I1005 08:02:18.360381 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:19 crc kubenswrapper[4935]: I1005 08:02:19.070602 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:19 crc kubenswrapper[4935]: I1005 08:02:19.124720 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fljbf"] Oct 05 08:02:21 crc kubenswrapper[4935]: I1005 08:02:21.030813 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fljbf" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerName="registry-server" containerID="cri-o://f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804" gracePeriod=2 Oct 05 08:02:21 crc kubenswrapper[4935]: I1005 08:02:21.939217 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.040422 4935 generic.go:334] "Generic (PLEG): container finished" podID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerID="f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804" exitCode=0 Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.040471 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fljbf" event={"ID":"591a2fb1-8a31-4a87-8f5d-26d111519c78","Type":"ContainerDied","Data":"f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804"} Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.040498 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fljbf" event={"ID":"591a2fb1-8a31-4a87-8f5d-26d111519c78","Type":"ContainerDied","Data":"470a5f7315536dfbc7a5ccc06a106c24da549a5ae95f3c97af6b1409d56b1f1a"} Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.040516 4935 scope.go:117] "RemoveContainer" containerID="f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.040530 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fljbf" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.057330 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-utilities\") pod \"591a2fb1-8a31-4a87-8f5d-26d111519c78\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.057480 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-catalog-content\") pod \"591a2fb1-8a31-4a87-8f5d-26d111519c78\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.057608 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfph5\" (UniqueName: \"kubernetes.io/projected/591a2fb1-8a31-4a87-8f5d-26d111519c78-kube-api-access-cfph5\") pod \"591a2fb1-8a31-4a87-8f5d-26d111519c78\" (UID: \"591a2fb1-8a31-4a87-8f5d-26d111519c78\") " Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.058137 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-utilities" (OuterVolumeSpecName: "utilities") pod "591a2fb1-8a31-4a87-8f5d-26d111519c78" (UID: "591a2fb1-8a31-4a87-8f5d-26d111519c78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.062599 4935 scope.go:117] "RemoveContainer" containerID="6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.063552 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/591a2fb1-8a31-4a87-8f5d-26d111519c78-kube-api-access-cfph5" (OuterVolumeSpecName: "kube-api-access-cfph5") pod "591a2fb1-8a31-4a87-8f5d-26d111519c78" (UID: "591a2fb1-8a31-4a87-8f5d-26d111519c78"). InnerVolumeSpecName "kube-api-access-cfph5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.097130 4935 scope.go:117] "RemoveContainer" containerID="9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.121220 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "591a2fb1-8a31-4a87-8f5d-26d111519c78" (UID: "591a2fb1-8a31-4a87-8f5d-26d111519c78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.124051 4935 scope.go:117] "RemoveContainer" containerID="f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804" Oct 05 08:02:22 crc kubenswrapper[4935]: E1005 08:02:22.124522 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804\": container with ID starting with f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804 not found: ID does not exist" containerID="f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.124556 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804"} err="failed to get container status \"f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804\": rpc error: code = NotFound desc = could not find container \"f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804\": container with ID starting with f948fb0e744fea416ee1f2a7db810bf97f4d3ca5465f71125b357e61a3d05804 not found: ID does not exist" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.124575 4935 scope.go:117] "RemoveContainer" containerID="6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8" Oct 05 08:02:22 crc kubenswrapper[4935]: E1005 08:02:22.124868 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8\": container with ID starting with 6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8 not found: ID does not exist" containerID="6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.124888 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8"} err="failed to get container status \"6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8\": rpc error: code = NotFound desc = could not find container \"6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8\": container with ID starting with 6e8046ab36e5cf3777a22700830ec970e146be20207f4792c740cc4c8a0fefe8 not found: ID does not exist" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.124926 4935 scope.go:117] "RemoveContainer" containerID="9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91" Oct 05 08:02:22 crc kubenswrapper[4935]: E1005 08:02:22.125226 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91\": container with ID starting with 9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91 not found: ID does not exist" containerID="9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.125246 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91"} err="failed to get container status \"9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91\": rpc error: code = NotFound desc = could not find container \"9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91\": container with ID starting with 9c41bfacdd4c1e1ba56451b59f2fc2ac6e2a293a9d8b43fab40d41b21512de91 not found: ID does not exist" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.159702 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.159757 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591a2fb1-8a31-4a87-8f5d-26d111519c78-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.159781 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfph5\" (UniqueName: \"kubernetes.io/projected/591a2fb1-8a31-4a87-8f5d-26d111519c78-kube-api-access-cfph5\") on node \"crc\" DevicePath \"\"" Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.368395 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fljbf"] Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.387209 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fljbf"] Oct 05 08:02:22 crc kubenswrapper[4935]: I1005 08:02:22.792186 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" path="/var/lib/kubelet/pods/591a2fb1-8a31-4a87-8f5d-26d111519c78/volumes" Oct 05 08:02:44 crc kubenswrapper[4935]: I1005 08:02:44.290110 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:02:44 crc kubenswrapper[4935]: I1005 08:02:44.290736 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.289318 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.289748 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.289795 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.290491 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.290547 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" gracePeriod=600 Oct 05 08:03:14 crc kubenswrapper[4935]: E1005 08:03:14.412619 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.480356 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" exitCode=0 Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.480402 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c"} Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.480443 4935 scope.go:117] "RemoveContainer" containerID="01e85d791e8730d0b4166d9cd3de9cf5c5a5f22d1c7dfc35c0721f631ba9149d" Oct 05 08:03:14 crc kubenswrapper[4935]: I1005 08:03:14.481173 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:03:14 crc kubenswrapper[4935]: E1005 08:03:14.481490 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:03:24 crc kubenswrapper[4935]: I1005 08:03:24.778096 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:03:24 crc kubenswrapper[4935]: E1005 08:03:24.779193 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:03:38 crc kubenswrapper[4935]: I1005 08:03:38.777669 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:03:38 crc kubenswrapper[4935]: E1005 08:03:38.778609 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:03:52 crc kubenswrapper[4935]: I1005 08:03:52.777296 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:03:52 crc kubenswrapper[4935]: E1005 08:03:52.778561 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:04:06 crc kubenswrapper[4935]: I1005 08:04:06.786744 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:04:06 crc kubenswrapper[4935]: E1005 08:04:06.787955 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:04:17 crc kubenswrapper[4935]: I1005 08:04:17.778357 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:04:17 crc kubenswrapper[4935]: E1005 08:04:17.779677 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:04:31 crc kubenswrapper[4935]: I1005 08:04:31.777694 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:04:31 crc kubenswrapper[4935]: E1005 08:04:31.778850 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:04:42 crc kubenswrapper[4935]: I1005 08:04:42.777312 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:04:42 crc kubenswrapper[4935]: E1005 08:04:42.779582 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:04:55 crc kubenswrapper[4935]: I1005 08:04:55.777665 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:04:55 crc kubenswrapper[4935]: E1005 08:04:55.778839 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:05:10 crc kubenswrapper[4935]: I1005 08:05:10.778419 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:05:10 crc kubenswrapper[4935]: E1005 08:05:10.779984 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:05:21 crc kubenswrapper[4935]: I1005 08:05:21.777269 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:05:21 crc kubenswrapper[4935]: E1005 08:05:21.778309 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:05:31 crc kubenswrapper[4935]: I1005 08:05:31.918269 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9792x"] Oct 05 08:05:31 crc kubenswrapper[4935]: E1005 08:05:31.919345 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerName="extract-content" Oct 05 08:05:31 crc kubenswrapper[4935]: I1005 08:05:31.919378 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerName="extract-content" Oct 05 08:05:31 crc kubenswrapper[4935]: E1005 08:05:31.919408 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerName="extract-utilities" Oct 05 08:05:31 crc kubenswrapper[4935]: I1005 08:05:31.919427 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerName="extract-utilities" Oct 05 08:05:31 crc kubenswrapper[4935]: E1005 08:05:31.919459 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerName="registry-server" Oct 05 08:05:31 crc kubenswrapper[4935]: I1005 08:05:31.919478 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerName="registry-server" Oct 05 08:05:31 crc kubenswrapper[4935]: I1005 08:05:31.919933 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="591a2fb1-8a31-4a87-8f5d-26d111519c78" containerName="registry-server" Oct 05 08:05:31 crc kubenswrapper[4935]: I1005 08:05:31.922602 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:31 crc kubenswrapper[4935]: I1005 08:05:31.933656 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9792x"] Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.086620 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzhh4\" (UniqueName: \"kubernetes.io/projected/16b7b5cb-3311-49e7-b868-d2af44d5403b-kube-api-access-qzhh4\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.086677 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-utilities\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.086717 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-catalog-content\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.188198 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzhh4\" (UniqueName: \"kubernetes.io/projected/16b7b5cb-3311-49e7-b868-d2af44d5403b-kube-api-access-qzhh4\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.188307 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-utilities\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.188357 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-catalog-content\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.188975 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-catalog-content\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.189037 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-utilities\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.361148 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzhh4\" (UniqueName: \"kubernetes.io/projected/16b7b5cb-3311-49e7-b868-d2af44d5403b-kube-api-access-qzhh4\") pod \"redhat-marketplace-9792x\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.556134 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:32 crc kubenswrapper[4935]: I1005 08:05:32.776910 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:05:32 crc kubenswrapper[4935]: E1005 08:05:32.777410 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:05:33 crc kubenswrapper[4935]: I1005 08:05:33.011036 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9792x"] Oct 05 08:05:33 crc kubenswrapper[4935]: I1005 08:05:33.724553 4935 generic.go:334] "Generic (PLEG): container finished" podID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerID="527a42ff1627afc580908f6350fbdc45af6e5db1e750018c62715fb01ef6a07f" exitCode=0 Oct 05 08:05:33 crc kubenswrapper[4935]: I1005 08:05:33.724806 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9792x" event={"ID":"16b7b5cb-3311-49e7-b868-d2af44d5403b","Type":"ContainerDied","Data":"527a42ff1627afc580908f6350fbdc45af6e5db1e750018c62715fb01ef6a07f"} Oct 05 08:05:33 crc kubenswrapper[4935]: I1005 08:05:33.724963 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9792x" event={"ID":"16b7b5cb-3311-49e7-b868-d2af44d5403b","Type":"ContainerStarted","Data":"613126ebc5818bb1fb71a42eb955c2ff2e6dda933f1033a543cc65a0c9d5bde6"} Oct 05 08:05:33 crc kubenswrapper[4935]: I1005 08:05:33.729383 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:05:35 crc kubenswrapper[4935]: I1005 08:05:35.753149 4935 generic.go:334] "Generic (PLEG): container finished" podID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerID="11724d13ff6b678a4e28bfb9d2e5994c0e9334e86b9b3565d3b16a0f8d3b97af" exitCode=0 Oct 05 08:05:35 crc kubenswrapper[4935]: I1005 08:05:35.753696 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9792x" event={"ID":"16b7b5cb-3311-49e7-b868-d2af44d5403b","Type":"ContainerDied","Data":"11724d13ff6b678a4e28bfb9d2e5994c0e9334e86b9b3565d3b16a0f8d3b97af"} Oct 05 08:05:36 crc kubenswrapper[4935]: I1005 08:05:36.767870 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9792x" event={"ID":"16b7b5cb-3311-49e7-b868-d2af44d5403b","Type":"ContainerStarted","Data":"6c3c472e19e12044d53ea88516ca88467dc0e3a7286a3d5b2eee550425ae398f"} Oct 05 08:05:36 crc kubenswrapper[4935]: I1005 08:05:36.792429 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9792x" podStartSLOduration=3.267618426 podStartE2EDuration="5.792401988s" podCreationTimestamp="2025-10-05 08:05:31 +0000 UTC" firstStartedPulling="2025-10-05 08:05:33.728941511 +0000 UTC m=+4367.611568011" lastFinishedPulling="2025-10-05 08:05:36.253725103 +0000 UTC m=+4370.136351573" observedRunningTime="2025-10-05 08:05:36.791624077 +0000 UTC m=+4370.674250577" watchObservedRunningTime="2025-10-05 08:05:36.792401988 +0000 UTC m=+4370.675028488" Oct 05 08:05:42 crc kubenswrapper[4935]: I1005 08:05:42.556623 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:42 crc kubenswrapper[4935]: I1005 08:05:42.558413 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:42 crc kubenswrapper[4935]: I1005 08:05:42.633436 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:42 crc kubenswrapper[4935]: I1005 08:05:42.894942 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:42 crc kubenswrapper[4935]: I1005 08:05:42.940873 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9792x"] Oct 05 08:05:43 crc kubenswrapper[4935]: I1005 08:05:43.777958 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:05:43 crc kubenswrapper[4935]: E1005 08:05:43.778598 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:05:44 crc kubenswrapper[4935]: I1005 08:05:44.845497 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9792x" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerName="registry-server" containerID="cri-o://6c3c472e19e12044d53ea88516ca88467dc0e3a7286a3d5b2eee550425ae398f" gracePeriod=2 Oct 05 08:05:45 crc kubenswrapper[4935]: I1005 08:05:45.858512 4935 generic.go:334] "Generic (PLEG): container finished" podID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerID="6c3c472e19e12044d53ea88516ca88467dc0e3a7286a3d5b2eee550425ae398f" exitCode=0 Oct 05 08:05:45 crc kubenswrapper[4935]: I1005 08:05:45.858531 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9792x" event={"ID":"16b7b5cb-3311-49e7-b868-d2af44d5403b","Type":"ContainerDied","Data":"6c3c472e19e12044d53ea88516ca88467dc0e3a7286a3d5b2eee550425ae398f"} Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.038360 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.210209 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzhh4\" (UniqueName: \"kubernetes.io/projected/16b7b5cb-3311-49e7-b868-d2af44d5403b-kube-api-access-qzhh4\") pod \"16b7b5cb-3311-49e7-b868-d2af44d5403b\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.210650 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-catalog-content\") pod \"16b7b5cb-3311-49e7-b868-d2af44d5403b\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.210939 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-utilities\") pod \"16b7b5cb-3311-49e7-b868-d2af44d5403b\" (UID: \"16b7b5cb-3311-49e7-b868-d2af44d5403b\") " Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.211561 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-utilities" (OuterVolumeSpecName: "utilities") pod "16b7b5cb-3311-49e7-b868-d2af44d5403b" (UID: "16b7b5cb-3311-49e7-b868-d2af44d5403b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.227834 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16b7b5cb-3311-49e7-b868-d2af44d5403b" (UID: "16b7b5cb-3311-49e7-b868-d2af44d5403b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.260551 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16b7b5cb-3311-49e7-b868-d2af44d5403b-kube-api-access-qzhh4" (OuterVolumeSpecName: "kube-api-access-qzhh4") pod "16b7b5cb-3311-49e7-b868-d2af44d5403b" (UID: "16b7b5cb-3311-49e7-b868-d2af44d5403b"). InnerVolumeSpecName "kube-api-access-qzhh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.312641 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzhh4\" (UniqueName: \"kubernetes.io/projected/16b7b5cb-3311-49e7-b868-d2af44d5403b-kube-api-access-qzhh4\") on node \"crc\" DevicePath \"\"" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.312671 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.312681 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16b7b5cb-3311-49e7-b868-d2af44d5403b-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.875416 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9792x" event={"ID":"16b7b5cb-3311-49e7-b868-d2af44d5403b","Type":"ContainerDied","Data":"613126ebc5818bb1fb71a42eb955c2ff2e6dda933f1033a543cc65a0c9d5bde6"} Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.875498 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9792x" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.875988 4935 scope.go:117] "RemoveContainer" containerID="6c3c472e19e12044d53ea88516ca88467dc0e3a7286a3d5b2eee550425ae398f" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.907101 4935 scope.go:117] "RemoveContainer" containerID="11724d13ff6b678a4e28bfb9d2e5994c0e9334e86b9b3565d3b16a0f8d3b97af" Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.913018 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9792x"] Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.920096 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9792x"] Oct 05 08:05:46 crc kubenswrapper[4935]: I1005 08:05:46.935521 4935 scope.go:117] "RemoveContainer" containerID="527a42ff1627afc580908f6350fbdc45af6e5db1e750018c62715fb01ef6a07f" Oct 05 08:05:48 crc kubenswrapper[4935]: I1005 08:05:48.794347 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" path="/var/lib/kubelet/pods/16b7b5cb-3311-49e7-b868-d2af44d5403b/volumes" Oct 05 08:05:58 crc kubenswrapper[4935]: I1005 08:05:58.777067 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:05:58 crc kubenswrapper[4935]: E1005 08:05:58.779049 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:06:12 crc kubenswrapper[4935]: I1005 08:06:12.776980 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:06:12 crc kubenswrapper[4935]: E1005 08:06:12.777617 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:06:26 crc kubenswrapper[4935]: I1005 08:06:26.787111 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:06:26 crc kubenswrapper[4935]: E1005 08:06:26.788161 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:06:41 crc kubenswrapper[4935]: I1005 08:06:41.777842 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:06:41 crc kubenswrapper[4935]: E1005 08:06:41.778748 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:06:55 crc kubenswrapper[4935]: I1005 08:06:55.777954 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:06:55 crc kubenswrapper[4935]: E1005 08:06:55.779056 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:07:08 crc kubenswrapper[4935]: I1005 08:07:08.777496 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:07:08 crc kubenswrapper[4935]: E1005 08:07:08.778864 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:07:21 crc kubenswrapper[4935]: I1005 08:07:21.777244 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:07:21 crc kubenswrapper[4935]: E1005 08:07:21.777936 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:07:32 crc kubenswrapper[4935]: I1005 08:07:32.777445 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:07:32 crc kubenswrapper[4935]: E1005 08:07:32.779045 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:07:43 crc kubenswrapper[4935]: I1005 08:07:43.777129 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:07:43 crc kubenswrapper[4935]: E1005 08:07:43.779720 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:07:55 crc kubenswrapper[4935]: I1005 08:07:55.777237 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:07:55 crc kubenswrapper[4935]: E1005 08:07:55.778175 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:08:07 crc kubenswrapper[4935]: I1005 08:08:07.777317 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:08:07 crc kubenswrapper[4935]: E1005 08:08:07.779719 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:08:18 crc kubenswrapper[4935]: I1005 08:08:18.778082 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:08:19 crc kubenswrapper[4935]: I1005 08:08:19.243043 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"ac7a5fb2f46a8e56eda310a9316e6fab6b669e59662e507781fca1a68fe52602"} Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.084934 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cfhbw"] Oct 05 08:08:21 crc kubenswrapper[4935]: E1005 08:08:21.086978 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerName="extract-utilities" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.087075 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerName="extract-utilities" Oct 05 08:08:21 crc kubenswrapper[4935]: E1005 08:08:21.087156 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerName="extract-content" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.087219 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerName="extract-content" Oct 05 08:08:21 crc kubenswrapper[4935]: E1005 08:08:21.087298 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerName="registry-server" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.087354 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerName="registry-server" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.087537 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="16b7b5cb-3311-49e7-b868-d2af44d5403b" containerName="registry-server" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.088581 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.113948 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cfhbw"] Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.266380 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-catalog-content\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.266743 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-utilities\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.266768 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdrnw\" (UniqueName: \"kubernetes.io/projected/110d2267-6658-47a6-85a3-d3f8e4075d43-kube-api-access-vdrnw\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.367778 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdrnw\" (UniqueName: \"kubernetes.io/projected/110d2267-6658-47a6-85a3-d3f8e4075d43-kube-api-access-vdrnw\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.368231 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-catalog-content\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.368379 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-utilities\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.368690 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-catalog-content\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.368793 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-utilities\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.389144 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdrnw\" (UniqueName: \"kubernetes.io/projected/110d2267-6658-47a6-85a3-d3f8e4075d43-kube-api-access-vdrnw\") pod \"community-operators-cfhbw\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.407940 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:21 crc kubenswrapper[4935]: I1005 08:08:21.691624 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cfhbw"] Oct 05 08:08:22 crc kubenswrapper[4935]: I1005 08:08:22.264488 4935 generic.go:334] "Generic (PLEG): container finished" podID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerID="e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05" exitCode=0 Oct 05 08:08:22 crc kubenswrapper[4935]: I1005 08:08:22.264559 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfhbw" event={"ID":"110d2267-6658-47a6-85a3-d3f8e4075d43","Type":"ContainerDied","Data":"e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05"} Oct 05 08:08:22 crc kubenswrapper[4935]: I1005 08:08:22.264600 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfhbw" event={"ID":"110d2267-6658-47a6-85a3-d3f8e4075d43","Type":"ContainerStarted","Data":"6fd234461992bcb684a78a7a6dd8db7762fb521ab10f8282977c26d9e8b8212a"} Oct 05 08:08:23 crc kubenswrapper[4935]: I1005 08:08:23.273687 4935 generic.go:334] "Generic (PLEG): container finished" podID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerID="c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a" exitCode=0 Oct 05 08:08:23 crc kubenswrapper[4935]: I1005 08:08:23.273843 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfhbw" event={"ID":"110d2267-6658-47a6-85a3-d3f8e4075d43","Type":"ContainerDied","Data":"c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a"} Oct 05 08:08:24 crc kubenswrapper[4935]: I1005 08:08:24.285623 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfhbw" event={"ID":"110d2267-6658-47a6-85a3-d3f8e4075d43","Type":"ContainerStarted","Data":"c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064"} Oct 05 08:08:24 crc kubenswrapper[4935]: I1005 08:08:24.318220 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cfhbw" podStartSLOduration=1.819982258 podStartE2EDuration="3.318204147s" podCreationTimestamp="2025-10-05 08:08:21 +0000 UTC" firstStartedPulling="2025-10-05 08:08:22.266460995 +0000 UTC m=+4536.149087455" lastFinishedPulling="2025-10-05 08:08:23.764682844 +0000 UTC m=+4537.647309344" observedRunningTime="2025-10-05 08:08:24.313629548 +0000 UTC m=+4538.196256018" watchObservedRunningTime="2025-10-05 08:08:24.318204147 +0000 UTC m=+4538.200830607" Oct 05 08:08:31 crc kubenswrapper[4935]: I1005 08:08:31.408499 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:31 crc kubenswrapper[4935]: I1005 08:08:31.409061 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:31 crc kubenswrapper[4935]: I1005 08:08:31.468199 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:32 crc kubenswrapper[4935]: I1005 08:08:32.387348 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:32 crc kubenswrapper[4935]: I1005 08:08:32.434253 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cfhbw"] Oct 05 08:08:34 crc kubenswrapper[4935]: I1005 08:08:34.368105 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cfhbw" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerName="registry-server" containerID="cri-o://c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064" gracePeriod=2 Oct 05 08:08:34 crc kubenswrapper[4935]: I1005 08:08:34.842909 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.016148 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-utilities\") pod \"110d2267-6658-47a6-85a3-d3f8e4075d43\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.016572 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-catalog-content\") pod \"110d2267-6658-47a6-85a3-d3f8e4075d43\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.016608 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdrnw\" (UniqueName: \"kubernetes.io/projected/110d2267-6658-47a6-85a3-d3f8e4075d43-kube-api-access-vdrnw\") pod \"110d2267-6658-47a6-85a3-d3f8e4075d43\" (UID: \"110d2267-6658-47a6-85a3-d3f8e4075d43\") " Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.017569 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-utilities" (OuterVolumeSpecName: "utilities") pod "110d2267-6658-47a6-85a3-d3f8e4075d43" (UID: "110d2267-6658-47a6-85a3-d3f8e4075d43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.026313 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/110d2267-6658-47a6-85a3-d3f8e4075d43-kube-api-access-vdrnw" (OuterVolumeSpecName: "kube-api-access-vdrnw") pod "110d2267-6658-47a6-85a3-d3f8e4075d43" (UID: "110d2267-6658-47a6-85a3-d3f8e4075d43"). InnerVolumeSpecName "kube-api-access-vdrnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.080465 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "110d2267-6658-47a6-85a3-d3f8e4075d43" (UID: "110d2267-6658-47a6-85a3-d3f8e4075d43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.117999 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdrnw\" (UniqueName: \"kubernetes.io/projected/110d2267-6658-47a6-85a3-d3f8e4075d43-kube-api-access-vdrnw\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.118036 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.118051 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/110d2267-6658-47a6-85a3-d3f8e4075d43-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.376946 4935 generic.go:334] "Generic (PLEG): container finished" podID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerID="c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064" exitCode=0 Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.376991 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfhbw" event={"ID":"110d2267-6658-47a6-85a3-d3f8e4075d43","Type":"ContainerDied","Data":"c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064"} Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.377017 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfhbw" event={"ID":"110d2267-6658-47a6-85a3-d3f8e4075d43","Type":"ContainerDied","Data":"6fd234461992bcb684a78a7a6dd8db7762fb521ab10f8282977c26d9e8b8212a"} Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.377035 4935 scope.go:117] "RemoveContainer" containerID="c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.377037 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfhbw" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.402532 4935 scope.go:117] "RemoveContainer" containerID="c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.425959 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cfhbw"] Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.431843 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cfhbw"] Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.444087 4935 scope.go:117] "RemoveContainer" containerID="e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.464084 4935 scope.go:117] "RemoveContainer" containerID="c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064" Oct 05 08:08:35 crc kubenswrapper[4935]: E1005 08:08:35.464533 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064\": container with ID starting with c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064 not found: ID does not exist" containerID="c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.464572 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064"} err="failed to get container status \"c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064\": rpc error: code = NotFound desc = could not find container \"c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064\": container with ID starting with c819521b9a6746cf1a0fbc535d69fdc504b7fcfadaf3636d0b52412607159064 not found: ID does not exist" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.464598 4935 scope.go:117] "RemoveContainer" containerID="c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a" Oct 05 08:08:35 crc kubenswrapper[4935]: E1005 08:08:35.464924 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a\": container with ID starting with c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a not found: ID does not exist" containerID="c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.464951 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a"} err="failed to get container status \"c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a\": rpc error: code = NotFound desc = could not find container \"c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a\": container with ID starting with c86efbfd42b389f6f2aeb9c47082ef43cb95e5b6c655d24dd6824f9b6c2b6f3a not found: ID does not exist" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.464970 4935 scope.go:117] "RemoveContainer" containerID="e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05" Oct 05 08:08:35 crc kubenswrapper[4935]: E1005 08:08:35.465188 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05\": container with ID starting with e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05 not found: ID does not exist" containerID="e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05" Oct 05 08:08:35 crc kubenswrapper[4935]: I1005 08:08:35.465211 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05"} err="failed to get container status \"e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05\": rpc error: code = NotFound desc = could not find container \"e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05\": container with ID starting with e53b8081ab63aee6ce74d82b5f75fd4ff2ba75060bd488452acac72514a83c05 not found: ID does not exist" Oct 05 08:08:35 crc kubenswrapper[4935]: E1005 08:08:35.534910 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod110d2267_6658_47a6_85a3_d3f8e4075d43.slice/crio-6fd234461992bcb684a78a7a6dd8db7762fb521ab10f8282977c26d9e8b8212a\": RecentStats: unable to find data in memory cache]" Oct 05 08:08:36 crc kubenswrapper[4935]: I1005 08:08:36.799324 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" path="/var/lib/kubelet/pods/110d2267-6658-47a6-85a3-d3f8e4075d43/volumes" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.193240 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-94582"] Oct 05 08:10:36 crc kubenswrapper[4935]: E1005 08:10:36.194017 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerName="extract-content" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.194032 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerName="extract-content" Oct 05 08:10:36 crc kubenswrapper[4935]: E1005 08:10:36.194040 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerName="registry-server" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.194046 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerName="registry-server" Oct 05 08:10:36 crc kubenswrapper[4935]: E1005 08:10:36.194061 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerName="extract-utilities" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.194067 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerName="extract-utilities" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.194211 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="110d2267-6658-47a6-85a3-d3f8e4075d43" containerName="registry-server" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.195205 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.213345 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-94582"] Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.315211 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvccp\" (UniqueName: \"kubernetes.io/projected/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-kube-api-access-qvccp\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.315258 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-utilities\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.315433 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-catalog-content\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.417404 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvccp\" (UniqueName: \"kubernetes.io/projected/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-kube-api-access-qvccp\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.417459 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-utilities\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.417528 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-catalog-content\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.418048 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-catalog-content\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.418166 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-utilities\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.444182 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvccp\" (UniqueName: \"kubernetes.io/projected/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-kube-api-access-qvccp\") pod \"redhat-operators-94582\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.530689 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:36 crc kubenswrapper[4935]: I1005 08:10:36.967108 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-94582"] Oct 05 08:10:37 crc kubenswrapper[4935]: I1005 08:10:37.459792 4935 generic.go:334] "Generic (PLEG): container finished" podID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerID="0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75" exitCode=0 Oct 05 08:10:37 crc kubenswrapper[4935]: I1005 08:10:37.459876 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94582" event={"ID":"5bbf9206-31f7-4aef-be04-0d2cfb37dba5","Type":"ContainerDied","Data":"0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75"} Oct 05 08:10:37 crc kubenswrapper[4935]: I1005 08:10:37.460081 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94582" event={"ID":"5bbf9206-31f7-4aef-be04-0d2cfb37dba5","Type":"ContainerStarted","Data":"319da618fd51b781f6c1890924056336256f47f834e53432180ace7d01032d42"} Oct 05 08:10:37 crc kubenswrapper[4935]: I1005 08:10:37.462154 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:10:38 crc kubenswrapper[4935]: I1005 08:10:38.495227 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94582" event={"ID":"5bbf9206-31f7-4aef-be04-0d2cfb37dba5","Type":"ContainerStarted","Data":"9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44"} Oct 05 08:10:39 crc kubenswrapper[4935]: I1005 08:10:39.504747 4935 generic.go:334] "Generic (PLEG): container finished" podID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerID="9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44" exitCode=0 Oct 05 08:10:39 crc kubenswrapper[4935]: I1005 08:10:39.504969 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94582" event={"ID":"5bbf9206-31f7-4aef-be04-0d2cfb37dba5","Type":"ContainerDied","Data":"9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44"} Oct 05 08:10:41 crc kubenswrapper[4935]: I1005 08:10:41.523656 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94582" event={"ID":"5bbf9206-31f7-4aef-be04-0d2cfb37dba5","Type":"ContainerStarted","Data":"c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85"} Oct 05 08:10:41 crc kubenswrapper[4935]: I1005 08:10:41.552053 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-94582" podStartSLOduration=2.957265306 podStartE2EDuration="5.552023602s" podCreationTimestamp="2025-10-05 08:10:36 +0000 UTC" firstStartedPulling="2025-10-05 08:10:37.461941478 +0000 UTC m=+4671.344567938" lastFinishedPulling="2025-10-05 08:10:40.056699774 +0000 UTC m=+4673.939326234" observedRunningTime="2025-10-05 08:10:41.543551208 +0000 UTC m=+4675.426177668" watchObservedRunningTime="2025-10-05 08:10:41.552023602 +0000 UTC m=+4675.434650092" Oct 05 08:10:44 crc kubenswrapper[4935]: I1005 08:10:44.289673 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:10:44 crc kubenswrapper[4935]: I1005 08:10:44.290067 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:10:46 crc kubenswrapper[4935]: I1005 08:10:46.548971 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:46 crc kubenswrapper[4935]: I1005 08:10:46.549035 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:46 crc kubenswrapper[4935]: I1005 08:10:46.630814 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:46 crc kubenswrapper[4935]: I1005 08:10:46.702905 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:46 crc kubenswrapper[4935]: I1005 08:10:46.866214 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-94582"] Oct 05 08:10:48 crc kubenswrapper[4935]: I1005 08:10:48.583032 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-94582" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerName="registry-server" containerID="cri-o://c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85" gracePeriod=2 Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.069948 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.238360 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-utilities\") pod \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.238436 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-catalog-content\") pod \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.238547 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvccp\" (UniqueName: \"kubernetes.io/projected/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-kube-api-access-qvccp\") pod \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\" (UID: \"5bbf9206-31f7-4aef-be04-0d2cfb37dba5\") " Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.239376 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-utilities" (OuterVolumeSpecName: "utilities") pod "5bbf9206-31f7-4aef-be04-0d2cfb37dba5" (UID: "5bbf9206-31f7-4aef-be04-0d2cfb37dba5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.245275 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-kube-api-access-qvccp" (OuterVolumeSpecName: "kube-api-access-qvccp") pod "5bbf9206-31f7-4aef-be04-0d2cfb37dba5" (UID: "5bbf9206-31f7-4aef-be04-0d2cfb37dba5"). InnerVolumeSpecName "kube-api-access-qvccp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.340978 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvccp\" (UniqueName: \"kubernetes.io/projected/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-kube-api-access-qvccp\") on node \"crc\" DevicePath \"\"" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.341187 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.591995 4935 generic.go:334] "Generic (PLEG): container finished" podID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerID="c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85" exitCode=0 Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.592137 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94582" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.592185 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94582" event={"ID":"5bbf9206-31f7-4aef-be04-0d2cfb37dba5","Type":"ContainerDied","Data":"c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85"} Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.592412 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94582" event={"ID":"5bbf9206-31f7-4aef-be04-0d2cfb37dba5","Type":"ContainerDied","Data":"319da618fd51b781f6c1890924056336256f47f834e53432180ace7d01032d42"} Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.592452 4935 scope.go:117] "RemoveContainer" containerID="c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.617720 4935 scope.go:117] "RemoveContainer" containerID="9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.634101 4935 scope.go:117] "RemoveContainer" containerID="0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.693376 4935 scope.go:117] "RemoveContainer" containerID="c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85" Oct 05 08:10:49 crc kubenswrapper[4935]: E1005 08:10:49.694234 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85\": container with ID starting with c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85 not found: ID does not exist" containerID="c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.694290 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85"} err="failed to get container status \"c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85\": rpc error: code = NotFound desc = could not find container \"c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85\": container with ID starting with c3c83179de04f5d6c3382c8f443996c5cb632006f2b750e9637b97542679be85 not found: ID does not exist" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.694321 4935 scope.go:117] "RemoveContainer" containerID="9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44" Oct 05 08:10:49 crc kubenswrapper[4935]: E1005 08:10:49.694760 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44\": container with ID starting with 9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44 not found: ID does not exist" containerID="9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.694831 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44"} err="failed to get container status \"9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44\": rpc error: code = NotFound desc = could not find container \"9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44\": container with ID starting with 9cf354b9ffc110ef826c2483672dc49a3e2872c3824e8f4838f5b5ba63979f44 not found: ID does not exist" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.694861 4935 scope.go:117] "RemoveContainer" containerID="0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75" Oct 05 08:10:49 crc kubenswrapper[4935]: E1005 08:10:49.695306 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75\": container with ID starting with 0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75 not found: ID does not exist" containerID="0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75" Oct 05 08:10:49 crc kubenswrapper[4935]: I1005 08:10:49.695377 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75"} err="failed to get container status \"0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75\": rpc error: code = NotFound desc = could not find container \"0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75\": container with ID starting with 0f881e0ada6f44a20d59549dcbd01c2cf353434eac54e3dcb3f019e037210e75 not found: ID does not exist" Oct 05 08:10:50 crc kubenswrapper[4935]: I1005 08:10:50.101960 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bbf9206-31f7-4aef-be04-0d2cfb37dba5" (UID: "5bbf9206-31f7-4aef-be04-0d2cfb37dba5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:10:50 crc kubenswrapper[4935]: I1005 08:10:50.153225 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf9206-31f7-4aef-be04-0d2cfb37dba5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:10:50 crc kubenswrapper[4935]: I1005 08:10:50.228188 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-94582"] Oct 05 08:10:50 crc kubenswrapper[4935]: I1005 08:10:50.237259 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-94582"] Oct 05 08:10:50 crc kubenswrapper[4935]: I1005 08:10:50.785020 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" path="/var/lib/kubelet/pods/5bbf9206-31f7-4aef-be04-0d2cfb37dba5/volumes" Oct 05 08:11:14 crc kubenswrapper[4935]: I1005 08:11:14.289501 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:11:14 crc kubenswrapper[4935]: I1005 08:11:14.290085 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:11:44 crc kubenswrapper[4935]: I1005 08:11:44.289132 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:11:44 crc kubenswrapper[4935]: I1005 08:11:44.289940 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:11:44 crc kubenswrapper[4935]: I1005 08:11:44.290018 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:11:44 crc kubenswrapper[4935]: I1005 08:11:44.291051 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac7a5fb2f46a8e56eda310a9316e6fab6b669e59662e507781fca1a68fe52602"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:11:44 crc kubenswrapper[4935]: I1005 08:11:44.291172 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://ac7a5fb2f46a8e56eda310a9316e6fab6b669e59662e507781fca1a68fe52602" gracePeriod=600 Oct 05 08:11:45 crc kubenswrapper[4935]: I1005 08:11:45.056814 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="ac7a5fb2f46a8e56eda310a9316e6fab6b669e59662e507781fca1a68fe52602" exitCode=0 Oct 05 08:11:45 crc kubenswrapper[4935]: I1005 08:11:45.056915 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"ac7a5fb2f46a8e56eda310a9316e6fab6b669e59662e507781fca1a68fe52602"} Oct 05 08:11:45 crc kubenswrapper[4935]: I1005 08:11:45.057397 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6"} Oct 05 08:11:45 crc kubenswrapper[4935]: I1005 08:11:45.057427 4935 scope.go:117] "RemoveContainer" containerID="b1a76e67e523c1d3f585b3d1338aa6fb6d3083ac777df80f5d47924736d4486c" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.166962 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-826z9"] Oct 05 08:12:56 crc kubenswrapper[4935]: E1005 08:12:56.167854 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerName="registry-server" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.167869 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerName="registry-server" Oct 05 08:12:56 crc kubenswrapper[4935]: E1005 08:12:56.167906 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerName="extract-content" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.167915 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerName="extract-content" Oct 05 08:12:56 crc kubenswrapper[4935]: E1005 08:12:56.167937 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerName="extract-utilities" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.167945 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerName="extract-utilities" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.168131 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bbf9206-31f7-4aef-be04-0d2cfb37dba5" containerName="registry-server" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.169565 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.185252 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-826z9"] Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.312666 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-utilities\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.312742 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-catalog-content\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.312773 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpz4d\" (UniqueName: \"kubernetes.io/projected/d886938e-6109-4731-ad22-3aba47806a87-kube-api-access-jpz4d\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.414433 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-utilities\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.414536 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-catalog-content\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.414571 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpz4d\" (UniqueName: \"kubernetes.io/projected/d886938e-6109-4731-ad22-3aba47806a87-kube-api-access-jpz4d\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.415006 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-catalog-content\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.415064 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-utilities\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.437913 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpz4d\" (UniqueName: \"kubernetes.io/projected/d886938e-6109-4731-ad22-3aba47806a87-kube-api-access-jpz4d\") pod \"certified-operators-826z9\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.489775 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:12:56 crc kubenswrapper[4935]: I1005 08:12:56.768314 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-826z9"] Oct 05 08:12:57 crc kubenswrapper[4935]: I1005 08:12:57.708814 4935 generic.go:334] "Generic (PLEG): container finished" podID="d886938e-6109-4731-ad22-3aba47806a87" containerID="7d61a31d7ad1d4590d58193aa4ca1bf0f1abef715ba532c9939a7bd40ef89775" exitCode=0 Oct 05 08:12:57 crc kubenswrapper[4935]: I1005 08:12:57.708950 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-826z9" event={"ID":"d886938e-6109-4731-ad22-3aba47806a87","Type":"ContainerDied","Data":"7d61a31d7ad1d4590d58193aa4ca1bf0f1abef715ba532c9939a7bd40ef89775"} Oct 05 08:12:57 crc kubenswrapper[4935]: I1005 08:12:57.709193 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-826z9" event={"ID":"d886938e-6109-4731-ad22-3aba47806a87","Type":"ContainerStarted","Data":"932786b016b2ba4ce641046132b1058f0537be2a0ec7d41a1e6d691b13a3098a"} Oct 05 08:12:58 crc kubenswrapper[4935]: I1005 08:12:58.721269 4935 generic.go:334] "Generic (PLEG): container finished" podID="d886938e-6109-4731-ad22-3aba47806a87" containerID="b981ad3d5d9ac026b9d2fc87b14786ea317f804735f1e55bf4723e69c199737c" exitCode=0 Oct 05 08:12:58 crc kubenswrapper[4935]: I1005 08:12:58.721408 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-826z9" event={"ID":"d886938e-6109-4731-ad22-3aba47806a87","Type":"ContainerDied","Data":"b981ad3d5d9ac026b9d2fc87b14786ea317f804735f1e55bf4723e69c199737c"} Oct 05 08:12:59 crc kubenswrapper[4935]: I1005 08:12:59.733964 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-826z9" event={"ID":"d886938e-6109-4731-ad22-3aba47806a87","Type":"ContainerStarted","Data":"aa84107fa2b2fc973fb924f21acad0901b9c7c52466b5f6234705376bad6fa16"} Oct 05 08:12:59 crc kubenswrapper[4935]: I1005 08:12:59.766862 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-826z9" podStartSLOduration=2.296829433 podStartE2EDuration="3.766834667s" podCreationTimestamp="2025-10-05 08:12:56 +0000 UTC" firstStartedPulling="2025-10-05 08:12:57.711836853 +0000 UTC m=+4811.594463313" lastFinishedPulling="2025-10-05 08:12:59.181842087 +0000 UTC m=+4813.064468547" observedRunningTime="2025-10-05 08:12:59.757131941 +0000 UTC m=+4813.639758461" watchObservedRunningTime="2025-10-05 08:12:59.766834667 +0000 UTC m=+4813.649461167" Oct 05 08:13:06 crc kubenswrapper[4935]: I1005 08:13:06.490537 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:13:06 crc kubenswrapper[4935]: I1005 08:13:06.492217 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:13:06 crc kubenswrapper[4935]: I1005 08:13:06.600771 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:13:06 crc kubenswrapper[4935]: I1005 08:13:06.842183 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:13:06 crc kubenswrapper[4935]: I1005 08:13:06.892366 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-826z9"] Oct 05 08:13:08 crc kubenswrapper[4935]: I1005 08:13:08.805618 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-826z9" podUID="d886938e-6109-4731-ad22-3aba47806a87" containerName="registry-server" containerID="cri-o://aa84107fa2b2fc973fb924f21acad0901b9c7c52466b5f6234705376bad6fa16" gracePeriod=2 Oct 05 08:13:09 crc kubenswrapper[4935]: I1005 08:13:09.816957 4935 generic.go:334] "Generic (PLEG): container finished" podID="d886938e-6109-4731-ad22-3aba47806a87" containerID="aa84107fa2b2fc973fb924f21acad0901b9c7c52466b5f6234705376bad6fa16" exitCode=0 Oct 05 08:13:09 crc kubenswrapper[4935]: I1005 08:13:09.816994 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-826z9" event={"ID":"d886938e-6109-4731-ad22-3aba47806a87","Type":"ContainerDied","Data":"aa84107fa2b2fc973fb924f21acad0901b9c7c52466b5f6234705376bad6fa16"} Oct 05 08:13:09 crc kubenswrapper[4935]: I1005 08:13:09.919480 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.109318 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpz4d\" (UniqueName: \"kubernetes.io/projected/d886938e-6109-4731-ad22-3aba47806a87-kube-api-access-jpz4d\") pod \"d886938e-6109-4731-ad22-3aba47806a87\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.109402 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-catalog-content\") pod \"d886938e-6109-4731-ad22-3aba47806a87\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.109512 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-utilities\") pod \"d886938e-6109-4731-ad22-3aba47806a87\" (UID: \"d886938e-6109-4731-ad22-3aba47806a87\") " Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.110939 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-utilities" (OuterVolumeSpecName: "utilities") pod "d886938e-6109-4731-ad22-3aba47806a87" (UID: "d886938e-6109-4731-ad22-3aba47806a87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.128210 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d886938e-6109-4731-ad22-3aba47806a87-kube-api-access-jpz4d" (OuterVolumeSpecName: "kube-api-access-jpz4d") pod "d886938e-6109-4731-ad22-3aba47806a87" (UID: "d886938e-6109-4731-ad22-3aba47806a87"). InnerVolumeSpecName "kube-api-access-jpz4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.151226 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d886938e-6109-4731-ad22-3aba47806a87" (UID: "d886938e-6109-4731-ad22-3aba47806a87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.211734 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpz4d\" (UniqueName: \"kubernetes.io/projected/d886938e-6109-4731-ad22-3aba47806a87-kube-api-access-jpz4d\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.211781 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.211792 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d886938e-6109-4731-ad22-3aba47806a87-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.832130 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-826z9" event={"ID":"d886938e-6109-4731-ad22-3aba47806a87","Type":"ContainerDied","Data":"932786b016b2ba4ce641046132b1058f0537be2a0ec7d41a1e6d691b13a3098a"} Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.832212 4935 scope.go:117] "RemoveContainer" containerID="aa84107fa2b2fc973fb924f21acad0901b9c7c52466b5f6234705376bad6fa16" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.832320 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-826z9" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.866335 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-826z9"] Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.878504 4935 scope.go:117] "RemoveContainer" containerID="b981ad3d5d9ac026b9d2fc87b14786ea317f804735f1e55bf4723e69c199737c" Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.888551 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-826z9"] Oct 05 08:13:10 crc kubenswrapper[4935]: I1005 08:13:10.911407 4935 scope.go:117] "RemoveContainer" containerID="7d61a31d7ad1d4590d58193aa4ca1bf0f1abef715ba532c9939a7bd40ef89775" Oct 05 08:13:12 crc kubenswrapper[4935]: I1005 08:13:12.790664 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d886938e-6109-4731-ad22-3aba47806a87" path="/var/lib/kubelet/pods/d886938e-6109-4731-ad22-3aba47806a87/volumes" Oct 05 08:13:44 crc kubenswrapper[4935]: I1005 08:13:44.289665 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:13:44 crc kubenswrapper[4935]: I1005 08:13:44.290569 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:14:14 crc kubenswrapper[4935]: I1005 08:14:14.289556 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:14:14 crc kubenswrapper[4935]: I1005 08:14:14.290205 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.288936 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.289309 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.289347 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.289827 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.289874 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" gracePeriod=600 Oct 05 08:14:44 crc kubenswrapper[4935]: E1005 08:14:44.420088 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.625999 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" exitCode=0 Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.626096 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6"} Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.626676 4935 scope.go:117] "RemoveContainer" containerID="ac7a5fb2f46a8e56eda310a9316e6fab6b669e59662e507781fca1a68fe52602" Oct 05 08:14:44 crc kubenswrapper[4935]: I1005 08:14:44.627514 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:14:44 crc kubenswrapper[4935]: E1005 08:14:44.627856 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:14:57 crc kubenswrapper[4935]: I1005 08:14:57.778650 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:14:57 crc kubenswrapper[4935]: E1005 08:14:57.779658 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.168237 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx"] Oct 05 08:15:00 crc kubenswrapper[4935]: E1005 08:15:00.169248 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d886938e-6109-4731-ad22-3aba47806a87" containerName="extract-utilities" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.169277 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d886938e-6109-4731-ad22-3aba47806a87" containerName="extract-utilities" Oct 05 08:15:00 crc kubenswrapper[4935]: E1005 08:15:00.169318 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d886938e-6109-4731-ad22-3aba47806a87" containerName="registry-server" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.169335 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d886938e-6109-4731-ad22-3aba47806a87" containerName="registry-server" Oct 05 08:15:00 crc kubenswrapper[4935]: E1005 08:15:00.169356 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d886938e-6109-4731-ad22-3aba47806a87" containerName="extract-content" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.169373 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d886938e-6109-4731-ad22-3aba47806a87" containerName="extract-content" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.169794 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d886938e-6109-4731-ad22-3aba47806a87" containerName="registry-server" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.171087 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.174237 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.174611 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.181551 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx"] Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.353513 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38c72e22-c292-4f2a-99ac-810745491cf4-secret-volume\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.353584 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38c72e22-c292-4f2a-99ac-810745491cf4-config-volume\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.353770 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w42sz\" (UniqueName: \"kubernetes.io/projected/38c72e22-c292-4f2a-99ac-810745491cf4-kube-api-access-w42sz\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.455981 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w42sz\" (UniqueName: \"kubernetes.io/projected/38c72e22-c292-4f2a-99ac-810745491cf4-kube-api-access-w42sz\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.456201 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38c72e22-c292-4f2a-99ac-810745491cf4-secret-volume\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.456261 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38c72e22-c292-4f2a-99ac-810745491cf4-config-volume\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.457735 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38c72e22-c292-4f2a-99ac-810745491cf4-config-volume\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.470418 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38c72e22-c292-4f2a-99ac-810745491cf4-secret-volume\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.476613 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w42sz\" (UniqueName: \"kubernetes.io/projected/38c72e22-c292-4f2a-99ac-810745491cf4-kube-api-access-w42sz\") pod \"collect-profiles-29327535-7xjqx\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.500192 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:00 crc kubenswrapper[4935]: I1005 08:15:00.946798 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx"] Oct 05 08:15:01 crc kubenswrapper[4935]: I1005 08:15:01.781471 4935 generic.go:334] "Generic (PLEG): container finished" podID="38c72e22-c292-4f2a-99ac-810745491cf4" containerID="2a3c4d5d5b222b0fd1149801e3d85fa52afa7225192c25dabe55b80f02545549" exitCode=0 Oct 05 08:15:01 crc kubenswrapper[4935]: I1005 08:15:01.781577 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" event={"ID":"38c72e22-c292-4f2a-99ac-810745491cf4","Type":"ContainerDied","Data":"2a3c4d5d5b222b0fd1149801e3d85fa52afa7225192c25dabe55b80f02545549"} Oct 05 08:15:01 crc kubenswrapper[4935]: I1005 08:15:01.782018 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" event={"ID":"38c72e22-c292-4f2a-99ac-810745491cf4","Type":"ContainerStarted","Data":"d7bfc54f750709b5a368dbb99f8259ad6b1c35bad441fa4b299d83897786fe07"} Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.167093 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.300794 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38c72e22-c292-4f2a-99ac-810745491cf4-config-volume\") pod \"38c72e22-c292-4f2a-99ac-810745491cf4\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.301162 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w42sz\" (UniqueName: \"kubernetes.io/projected/38c72e22-c292-4f2a-99ac-810745491cf4-kube-api-access-w42sz\") pod \"38c72e22-c292-4f2a-99ac-810745491cf4\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.301216 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38c72e22-c292-4f2a-99ac-810745491cf4-secret-volume\") pod \"38c72e22-c292-4f2a-99ac-810745491cf4\" (UID: \"38c72e22-c292-4f2a-99ac-810745491cf4\") " Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.302175 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38c72e22-c292-4f2a-99ac-810745491cf4-config-volume" (OuterVolumeSpecName: "config-volume") pod "38c72e22-c292-4f2a-99ac-810745491cf4" (UID: "38c72e22-c292-4f2a-99ac-810745491cf4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.307961 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38c72e22-c292-4f2a-99ac-810745491cf4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "38c72e22-c292-4f2a-99ac-810745491cf4" (UID: "38c72e22-c292-4f2a-99ac-810745491cf4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.308089 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38c72e22-c292-4f2a-99ac-810745491cf4-kube-api-access-w42sz" (OuterVolumeSpecName: "kube-api-access-w42sz") pod "38c72e22-c292-4f2a-99ac-810745491cf4" (UID: "38c72e22-c292-4f2a-99ac-810745491cf4"). InnerVolumeSpecName "kube-api-access-w42sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.402886 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38c72e22-c292-4f2a-99ac-810745491cf4-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.402940 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w42sz\" (UniqueName: \"kubernetes.io/projected/38c72e22-c292-4f2a-99ac-810745491cf4-kube-api-access-w42sz\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.402954 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38c72e22-c292-4f2a-99ac-810745491cf4-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.798989 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" event={"ID":"38c72e22-c292-4f2a-99ac-810745491cf4","Type":"ContainerDied","Data":"d7bfc54f750709b5a368dbb99f8259ad6b1c35bad441fa4b299d83897786fe07"} Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.799064 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7bfc54f750709b5a368dbb99f8259ad6b1c35bad441fa4b299d83897786fe07" Oct 05 08:15:03 crc kubenswrapper[4935]: I1005 08:15:03.799080 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx" Oct 05 08:15:04 crc kubenswrapper[4935]: I1005 08:15:04.249629 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6"] Oct 05 08:15:04 crc kubenswrapper[4935]: I1005 08:15:04.255282 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-dgnh6"] Oct 05 08:15:04 crc kubenswrapper[4935]: I1005 08:15:04.785923 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03cd4b1b-dcaf-4c59-ba06-b8c11b061473" path="/var/lib/kubelet/pods/03cd4b1b-dcaf-4c59-ba06-b8c11b061473/volumes" Oct 05 08:15:11 crc kubenswrapper[4935]: I1005 08:15:11.778053 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:15:11 crc kubenswrapper[4935]: E1005 08:15:11.778762 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:15:26 crc kubenswrapper[4935]: I1005 08:15:26.777189 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:15:26 crc kubenswrapper[4935]: E1005 08:15:26.778159 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:15:37 crc kubenswrapper[4935]: I1005 08:15:37.777286 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:15:37 crc kubenswrapper[4935]: E1005 08:15:37.778501 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:15:50 crc kubenswrapper[4935]: I1005 08:15:50.778218 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:15:50 crc kubenswrapper[4935]: E1005 08:15:50.781055 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:15:55 crc kubenswrapper[4935]: I1005 08:15:55.927732 4935 scope.go:117] "RemoveContainer" containerID="24bc98efcf1c1d38bdba0dd62124c501ac653ecf08f63074bb6714380331abcf" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.032445 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w4stl"] Oct 05 08:15:59 crc kubenswrapper[4935]: E1005 08:15:59.033139 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c72e22-c292-4f2a-99ac-810745491cf4" containerName="collect-profiles" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.033157 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c72e22-c292-4f2a-99ac-810745491cf4" containerName="collect-profiles" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.033328 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c72e22-c292-4f2a-99ac-810745491cf4" containerName="collect-profiles" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.034850 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.064632 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4stl"] Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.151095 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m295\" (UniqueName: \"kubernetes.io/projected/e3a1301b-d71f-45e5-b1ef-991f062b25a3-kube-api-access-8m295\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.151353 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-utilities\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.151518 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-catalog-content\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.253596 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m295\" (UniqueName: \"kubernetes.io/projected/e3a1301b-d71f-45e5-b1ef-991f062b25a3-kube-api-access-8m295\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.253731 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-utilities\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.253831 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-catalog-content\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.254387 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-catalog-content\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.254783 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-utilities\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.291866 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m295\" (UniqueName: \"kubernetes.io/projected/e3a1301b-d71f-45e5-b1ef-991f062b25a3-kube-api-access-8m295\") pod \"redhat-marketplace-w4stl\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.359904 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:15:59 crc kubenswrapper[4935]: I1005 08:15:59.788837 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4stl"] Oct 05 08:16:00 crc kubenswrapper[4935]: I1005 08:16:00.245023 4935 generic.go:334] "Generic (PLEG): container finished" podID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerID="bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a" exitCode=0 Oct 05 08:16:00 crc kubenswrapper[4935]: I1005 08:16:00.245071 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4stl" event={"ID":"e3a1301b-d71f-45e5-b1ef-991f062b25a3","Type":"ContainerDied","Data":"bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a"} Oct 05 08:16:00 crc kubenswrapper[4935]: I1005 08:16:00.245099 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4stl" event={"ID":"e3a1301b-d71f-45e5-b1ef-991f062b25a3","Type":"ContainerStarted","Data":"59c7fd14c6f241d1586e10ba872f44f43352eadcdfe06d495ce1ed77ff8bf44b"} Oct 05 08:16:00 crc kubenswrapper[4935]: I1005 08:16:00.248247 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:16:02 crc kubenswrapper[4935]: I1005 08:16:02.262461 4935 generic.go:334] "Generic (PLEG): container finished" podID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerID="feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df" exitCode=0 Oct 05 08:16:02 crc kubenswrapper[4935]: I1005 08:16:02.262569 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4stl" event={"ID":"e3a1301b-d71f-45e5-b1ef-991f062b25a3","Type":"ContainerDied","Data":"feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df"} Oct 05 08:16:03 crc kubenswrapper[4935]: I1005 08:16:03.280177 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4stl" event={"ID":"e3a1301b-d71f-45e5-b1ef-991f062b25a3","Type":"ContainerStarted","Data":"10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299"} Oct 05 08:16:03 crc kubenswrapper[4935]: I1005 08:16:03.307237 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w4stl" podStartSLOduration=2.851791157 podStartE2EDuration="5.307202233s" podCreationTimestamp="2025-10-05 08:15:58 +0000 UTC" firstStartedPulling="2025-10-05 08:16:00.247978751 +0000 UTC m=+4994.130605211" lastFinishedPulling="2025-10-05 08:16:02.703389817 +0000 UTC m=+4996.586016287" observedRunningTime="2025-10-05 08:16:03.303632668 +0000 UTC m=+4997.186259148" watchObservedRunningTime="2025-10-05 08:16:03.307202233 +0000 UTC m=+4997.189828693" Oct 05 08:16:03 crc kubenswrapper[4935]: I1005 08:16:03.777668 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:16:03 crc kubenswrapper[4935]: E1005 08:16:03.778002 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:16:09 crc kubenswrapper[4935]: I1005 08:16:09.361028 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:16:09 crc kubenswrapper[4935]: I1005 08:16:09.361920 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:16:09 crc kubenswrapper[4935]: I1005 08:16:09.426571 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:16:10 crc kubenswrapper[4935]: I1005 08:16:10.410161 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:16:10 crc kubenswrapper[4935]: I1005 08:16:10.466295 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4stl"] Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.351026 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w4stl" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerName="registry-server" containerID="cri-o://10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299" gracePeriod=2 Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.776832 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.863059 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-catalog-content\") pod \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.863141 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-utilities\") pod \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.863288 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m295\" (UniqueName: \"kubernetes.io/projected/e3a1301b-d71f-45e5-b1ef-991f062b25a3-kube-api-access-8m295\") pod \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\" (UID: \"e3a1301b-d71f-45e5-b1ef-991f062b25a3\") " Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.866338 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-utilities" (OuterVolumeSpecName: "utilities") pod "e3a1301b-d71f-45e5-b1ef-991f062b25a3" (UID: "e3a1301b-d71f-45e5-b1ef-991f062b25a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.877299 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3a1301b-d71f-45e5-b1ef-991f062b25a3-kube-api-access-8m295" (OuterVolumeSpecName: "kube-api-access-8m295") pod "e3a1301b-d71f-45e5-b1ef-991f062b25a3" (UID: "e3a1301b-d71f-45e5-b1ef-991f062b25a3"). InnerVolumeSpecName "kube-api-access-8m295". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.878605 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3a1301b-d71f-45e5-b1ef-991f062b25a3" (UID: "e3a1301b-d71f-45e5-b1ef-991f062b25a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.965526 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m295\" (UniqueName: \"kubernetes.io/projected/e3a1301b-d71f-45e5-b1ef-991f062b25a3-kube-api-access-8m295\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.965571 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:12 crc kubenswrapper[4935]: I1005 08:16:12.965585 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a1301b-d71f-45e5-b1ef-991f062b25a3-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.375526 4935 generic.go:334] "Generic (PLEG): container finished" podID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerID="10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299" exitCode=0 Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.375611 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4stl" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.375612 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4stl" event={"ID":"e3a1301b-d71f-45e5-b1ef-991f062b25a3","Type":"ContainerDied","Data":"10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299"} Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.375680 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4stl" event={"ID":"e3a1301b-d71f-45e5-b1ef-991f062b25a3","Type":"ContainerDied","Data":"59c7fd14c6f241d1586e10ba872f44f43352eadcdfe06d495ce1ed77ff8bf44b"} Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.375712 4935 scope.go:117] "RemoveContainer" containerID="10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.411814 4935 scope.go:117] "RemoveContainer" containerID="feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.429333 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4stl"] Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.444079 4935 scope.go:117] "RemoveContainer" containerID="bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.444076 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4stl"] Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.482104 4935 scope.go:117] "RemoveContainer" containerID="10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299" Oct 05 08:16:13 crc kubenswrapper[4935]: E1005 08:16:13.482599 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299\": container with ID starting with 10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299 not found: ID does not exist" containerID="10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.482683 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299"} err="failed to get container status \"10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299\": rpc error: code = NotFound desc = could not find container \"10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299\": container with ID starting with 10bc88cb0286f305200234dd1ae87f8a43e3e1384824b213d3b72ae5c73d1299 not found: ID does not exist" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.482720 4935 scope.go:117] "RemoveContainer" containerID="feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df" Oct 05 08:16:13 crc kubenswrapper[4935]: E1005 08:16:13.483247 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df\": container with ID starting with feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df not found: ID does not exist" containerID="feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.483313 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df"} err="failed to get container status \"feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df\": rpc error: code = NotFound desc = could not find container \"feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df\": container with ID starting with feb3739048ecb0ba530ffa007aea88049653a0140d3a83453b452c9fa92527df not found: ID does not exist" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.483354 4935 scope.go:117] "RemoveContainer" containerID="bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a" Oct 05 08:16:13 crc kubenswrapper[4935]: E1005 08:16:13.483975 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a\": container with ID starting with bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a not found: ID does not exist" containerID="bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a" Oct 05 08:16:13 crc kubenswrapper[4935]: I1005 08:16:13.484015 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a"} err="failed to get container status \"bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a\": rpc error: code = NotFound desc = could not find container \"bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a\": container with ID starting with bc8167c0913b3dd503ed6a456617d7d46cca56fb86510bd373eae8211507b83a not found: ID does not exist" Oct 05 08:16:14 crc kubenswrapper[4935]: I1005 08:16:14.799171 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" path="/var/lib/kubelet/pods/e3a1301b-d71f-45e5-b1ef-991f062b25a3/volumes" Oct 05 08:16:17 crc kubenswrapper[4935]: I1005 08:16:17.776940 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:16:17 crc kubenswrapper[4935]: E1005 08:16:17.777752 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:16:28 crc kubenswrapper[4935]: I1005 08:16:28.778222 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:16:28 crc kubenswrapper[4935]: E1005 08:16:28.779460 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:16:43 crc kubenswrapper[4935]: I1005 08:16:43.777323 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:16:43 crc kubenswrapper[4935]: E1005 08:16:43.778274 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:16:56 crc kubenswrapper[4935]: I1005 08:16:56.783377 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:16:56 crc kubenswrapper[4935]: E1005 08:16:56.784084 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:17:10 crc kubenswrapper[4935]: I1005 08:17:10.777727 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:17:10 crc kubenswrapper[4935]: E1005 08:17:10.778514 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:17:24 crc kubenswrapper[4935]: I1005 08:17:24.777299 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:17:24 crc kubenswrapper[4935]: E1005 08:17:24.778375 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:17:35 crc kubenswrapper[4935]: I1005 08:17:35.777483 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:17:35 crc kubenswrapper[4935]: E1005 08:17:35.778560 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:17:46 crc kubenswrapper[4935]: I1005 08:17:46.785967 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:17:46 crc kubenswrapper[4935]: E1005 08:17:46.787352 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:18:01 crc kubenswrapper[4935]: I1005 08:18:01.776702 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:18:01 crc kubenswrapper[4935]: E1005 08:18:01.777309 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:18:14 crc kubenswrapper[4935]: I1005 08:18:14.776809 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:18:14 crc kubenswrapper[4935]: E1005 08:18:14.778109 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:18:29 crc kubenswrapper[4935]: I1005 08:18:29.777735 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:18:29 crc kubenswrapper[4935]: E1005 08:18:29.778707 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:18:44 crc kubenswrapper[4935]: I1005 08:18:44.777483 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:18:44 crc kubenswrapper[4935]: E1005 08:18:44.778920 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:18:59 crc kubenswrapper[4935]: I1005 08:18:59.777763 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:18:59 crc kubenswrapper[4935]: E1005 08:18:59.778755 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:19:14 crc kubenswrapper[4935]: I1005 08:19:14.778739 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:19:14 crc kubenswrapper[4935]: E1005 08:19:14.781242 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:19:27 crc kubenswrapper[4935]: I1005 08:19:27.777868 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:19:27 crc kubenswrapper[4935]: E1005 08:19:27.779014 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:19:39 crc kubenswrapper[4935]: I1005 08:19:39.777629 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:19:39 crc kubenswrapper[4935]: E1005 08:19:39.778335 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:19:50 crc kubenswrapper[4935]: I1005 08:19:50.777741 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:19:51 crc kubenswrapper[4935]: I1005 08:19:51.305080 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"ed3a0b66dfca5c0c324f5e00aad5979dd80a4661554edeae46ce1aa1c45c00be"} Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.607251 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5fmqf"] Oct 05 08:21:11 crc kubenswrapper[4935]: E1005 08:21:11.608312 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerName="extract-utilities" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.608332 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerName="extract-utilities" Oct 05 08:21:11 crc kubenswrapper[4935]: E1005 08:21:11.608361 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerName="registry-server" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.608371 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerName="registry-server" Oct 05 08:21:11 crc kubenswrapper[4935]: E1005 08:21:11.608401 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerName="extract-content" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.608414 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerName="extract-content" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.608656 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3a1301b-d71f-45e5-b1ef-991f062b25a3" containerName="registry-server" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.611071 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.633707 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5fmqf"] Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.728467 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-utilities\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.729137 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chwsl\" (UniqueName: \"kubernetes.io/projected/c10694e4-10b3-4199-9b84-56b08821252e-kube-api-access-chwsl\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.729364 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-catalog-content\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.831238 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-utilities\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.831303 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chwsl\" (UniqueName: \"kubernetes.io/projected/c10694e4-10b3-4199-9b84-56b08821252e-kube-api-access-chwsl\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.831349 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-catalog-content\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.831786 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-utilities\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.831830 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-catalog-content\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.850454 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chwsl\" (UniqueName: \"kubernetes.io/projected/c10694e4-10b3-4199-9b84-56b08821252e-kube-api-access-chwsl\") pod \"redhat-operators-5fmqf\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:11 crc kubenswrapper[4935]: I1005 08:21:11.980203 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:12 crc kubenswrapper[4935]: I1005 08:21:12.421742 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5fmqf"] Oct 05 08:21:13 crc kubenswrapper[4935]: I1005 08:21:13.015405 4935 generic.go:334] "Generic (PLEG): container finished" podID="c10694e4-10b3-4199-9b84-56b08821252e" containerID="4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5" exitCode=0 Oct 05 08:21:13 crc kubenswrapper[4935]: I1005 08:21:13.015455 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmqf" event={"ID":"c10694e4-10b3-4199-9b84-56b08821252e","Type":"ContainerDied","Data":"4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5"} Oct 05 08:21:13 crc kubenswrapper[4935]: I1005 08:21:13.015487 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmqf" event={"ID":"c10694e4-10b3-4199-9b84-56b08821252e","Type":"ContainerStarted","Data":"633b8abf39d2b4f322fb1abe4f56c35af0df79cdaf85ad32a1bbd1cd4fda2e05"} Oct 05 08:21:13 crc kubenswrapper[4935]: I1005 08:21:13.016932 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:21:14 crc kubenswrapper[4935]: I1005 08:21:14.026564 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmqf" event={"ID":"c10694e4-10b3-4199-9b84-56b08821252e","Type":"ContainerStarted","Data":"779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd"} Oct 05 08:21:15 crc kubenswrapper[4935]: I1005 08:21:15.034114 4935 generic.go:334] "Generic (PLEG): container finished" podID="c10694e4-10b3-4199-9b84-56b08821252e" containerID="779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd" exitCode=0 Oct 05 08:21:15 crc kubenswrapper[4935]: I1005 08:21:15.034218 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmqf" event={"ID":"c10694e4-10b3-4199-9b84-56b08821252e","Type":"ContainerDied","Data":"779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd"} Oct 05 08:21:16 crc kubenswrapper[4935]: I1005 08:21:16.048488 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmqf" event={"ID":"c10694e4-10b3-4199-9b84-56b08821252e","Type":"ContainerStarted","Data":"e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0"} Oct 05 08:21:16 crc kubenswrapper[4935]: I1005 08:21:16.080067 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5fmqf" podStartSLOduration=2.543656949 podStartE2EDuration="5.080041749s" podCreationTimestamp="2025-10-05 08:21:11 +0000 UTC" firstStartedPulling="2025-10-05 08:21:13.016637119 +0000 UTC m=+5306.899263579" lastFinishedPulling="2025-10-05 08:21:15.553021919 +0000 UTC m=+5309.435648379" observedRunningTime="2025-10-05 08:21:16.073932658 +0000 UTC m=+5309.956559148" watchObservedRunningTime="2025-10-05 08:21:16.080041749 +0000 UTC m=+5309.962668229" Oct 05 08:21:21 crc kubenswrapper[4935]: I1005 08:21:21.981454 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:21 crc kubenswrapper[4935]: I1005 08:21:21.982239 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:22 crc kubenswrapper[4935]: I1005 08:21:22.048547 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:22 crc kubenswrapper[4935]: I1005 08:21:22.157239 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:22 crc kubenswrapper[4935]: I1005 08:21:22.303225 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5fmqf"] Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.107799 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5fmqf" podUID="c10694e4-10b3-4199-9b84-56b08821252e" containerName="registry-server" containerID="cri-o://e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0" gracePeriod=2 Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.561500 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.616345 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-utilities\") pod \"c10694e4-10b3-4199-9b84-56b08821252e\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.616427 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-catalog-content\") pod \"c10694e4-10b3-4199-9b84-56b08821252e\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.616510 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chwsl\" (UniqueName: \"kubernetes.io/projected/c10694e4-10b3-4199-9b84-56b08821252e-kube-api-access-chwsl\") pod \"c10694e4-10b3-4199-9b84-56b08821252e\" (UID: \"c10694e4-10b3-4199-9b84-56b08821252e\") " Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.618087 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-utilities" (OuterVolumeSpecName: "utilities") pod "c10694e4-10b3-4199-9b84-56b08821252e" (UID: "c10694e4-10b3-4199-9b84-56b08821252e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.627699 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c10694e4-10b3-4199-9b84-56b08821252e-kube-api-access-chwsl" (OuterVolumeSpecName: "kube-api-access-chwsl") pod "c10694e4-10b3-4199-9b84-56b08821252e" (UID: "c10694e4-10b3-4199-9b84-56b08821252e"). InnerVolumeSpecName "kube-api-access-chwsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.704996 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c10694e4-10b3-4199-9b84-56b08821252e" (UID: "c10694e4-10b3-4199-9b84-56b08821252e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.717847 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.717876 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10694e4-10b3-4199-9b84-56b08821252e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:24 crc kubenswrapper[4935]: I1005 08:21:24.717904 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chwsl\" (UniqueName: \"kubernetes.io/projected/c10694e4-10b3-4199-9b84-56b08821252e-kube-api-access-chwsl\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.122515 4935 generic.go:334] "Generic (PLEG): container finished" podID="c10694e4-10b3-4199-9b84-56b08821252e" containerID="e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0" exitCode=0 Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.122561 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmqf" event={"ID":"c10694e4-10b3-4199-9b84-56b08821252e","Type":"ContainerDied","Data":"e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0"} Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.122594 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmqf" event={"ID":"c10694e4-10b3-4199-9b84-56b08821252e","Type":"ContainerDied","Data":"633b8abf39d2b4f322fb1abe4f56c35af0df79cdaf85ad32a1bbd1cd4fda2e05"} Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.122618 4935 scope.go:117] "RemoveContainer" containerID="e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.122625 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fmqf" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.154607 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5fmqf"] Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.161795 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5fmqf"] Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.166234 4935 scope.go:117] "RemoveContainer" containerID="779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.196383 4935 scope.go:117] "RemoveContainer" containerID="4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.223445 4935 scope.go:117] "RemoveContainer" containerID="e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0" Oct 05 08:21:25 crc kubenswrapper[4935]: E1005 08:21:25.224110 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0\": container with ID starting with e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0 not found: ID does not exist" containerID="e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.224155 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0"} err="failed to get container status \"e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0\": rpc error: code = NotFound desc = could not find container \"e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0\": container with ID starting with e182910cb351c748fa14b16c4b061efb4d48bbfac14875e372b65bc0d50f77b0 not found: ID does not exist" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.224182 4935 scope.go:117] "RemoveContainer" containerID="779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd" Oct 05 08:21:25 crc kubenswrapper[4935]: E1005 08:21:25.224695 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd\": container with ID starting with 779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd not found: ID does not exist" containerID="779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.224738 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd"} err="failed to get container status \"779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd\": rpc error: code = NotFound desc = could not find container \"779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd\": container with ID starting with 779e43fff946f6bbb393809153bfcdb2d7924913b67b8537f085a78e6916facd not found: ID does not exist" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.224765 4935 scope.go:117] "RemoveContainer" containerID="4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5" Oct 05 08:21:25 crc kubenswrapper[4935]: E1005 08:21:25.225268 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5\": container with ID starting with 4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5 not found: ID does not exist" containerID="4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5" Oct 05 08:21:25 crc kubenswrapper[4935]: I1005 08:21:25.225302 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5"} err="failed to get container status \"4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5\": rpc error: code = NotFound desc = could not find container \"4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5\": container with ID starting with 4d68c89a512d76b9fbca7c6a9c3a3e3670d09bf3b96fd2e001501abd85b67cf5 not found: ID does not exist" Oct 05 08:21:26 crc kubenswrapper[4935]: I1005 08:21:26.794799 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c10694e4-10b3-4199-9b84-56b08821252e" path="/var/lib/kubelet/pods/c10694e4-10b3-4199-9b84-56b08821252e/volumes" Oct 05 08:22:14 crc kubenswrapper[4935]: I1005 08:22:14.289291 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:22:14 crc kubenswrapper[4935]: I1005 08:22:14.289868 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:22:44 crc kubenswrapper[4935]: I1005 08:22:44.289976 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:22:44 crc kubenswrapper[4935]: I1005 08:22:44.290555 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:23:14 crc kubenswrapper[4935]: I1005 08:23:14.289489 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:23:14 crc kubenswrapper[4935]: I1005 08:23:14.290672 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:23:14 crc kubenswrapper[4935]: I1005 08:23:14.290790 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:23:14 crc kubenswrapper[4935]: I1005 08:23:14.291879 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed3a0b66dfca5c0c324f5e00aad5979dd80a4661554edeae46ce1aa1c45c00be"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:23:14 crc kubenswrapper[4935]: I1005 08:23:14.292003 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://ed3a0b66dfca5c0c324f5e00aad5979dd80a4661554edeae46ce1aa1c45c00be" gracePeriod=600 Oct 05 08:23:15 crc kubenswrapper[4935]: I1005 08:23:15.134877 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="ed3a0b66dfca5c0c324f5e00aad5979dd80a4661554edeae46ce1aa1c45c00be" exitCode=0 Oct 05 08:23:15 crc kubenswrapper[4935]: I1005 08:23:15.134935 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"ed3a0b66dfca5c0c324f5e00aad5979dd80a4661554edeae46ce1aa1c45c00be"} Oct 05 08:23:15 crc kubenswrapper[4935]: I1005 08:23:15.135489 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484"} Oct 05 08:23:15 crc kubenswrapper[4935]: I1005 08:23:15.135510 4935 scope.go:117] "RemoveContainer" containerID="4df350331030399082f4a299c8bcaf997c3fd8e9cd85c4af7dfb2c9c27c7c8d6" Oct 05 08:23:27 crc kubenswrapper[4935]: I1005 08:23:27.907315 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-scpvw"] Oct 05 08:23:27 crc kubenswrapper[4935]: E1005 08:23:27.908650 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10694e4-10b3-4199-9b84-56b08821252e" containerName="extract-utilities" Oct 05 08:23:27 crc kubenswrapper[4935]: I1005 08:23:27.908683 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10694e4-10b3-4199-9b84-56b08821252e" containerName="extract-utilities" Oct 05 08:23:27 crc kubenswrapper[4935]: E1005 08:23:27.908744 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10694e4-10b3-4199-9b84-56b08821252e" containerName="registry-server" Oct 05 08:23:27 crc kubenswrapper[4935]: I1005 08:23:27.908762 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10694e4-10b3-4199-9b84-56b08821252e" containerName="registry-server" Oct 05 08:23:27 crc kubenswrapper[4935]: E1005 08:23:27.908796 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10694e4-10b3-4199-9b84-56b08821252e" containerName="extract-content" Oct 05 08:23:27 crc kubenswrapper[4935]: I1005 08:23:27.908812 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10694e4-10b3-4199-9b84-56b08821252e" containerName="extract-content" Oct 05 08:23:27 crc kubenswrapper[4935]: I1005 08:23:27.909207 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c10694e4-10b3-4199-9b84-56b08821252e" containerName="registry-server" Oct 05 08:23:27 crc kubenswrapper[4935]: I1005 08:23:27.911881 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:27 crc kubenswrapper[4935]: I1005 08:23:27.927454 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-scpvw"] Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.059736 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-catalog-content\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.059832 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtzt9\" (UniqueName: \"kubernetes.io/projected/1ce971f4-45ea-463f-82da-687f425ddfb6-kube-api-access-dtzt9\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.059865 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-utilities\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.160828 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-catalog-content\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.160904 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtzt9\" (UniqueName: \"kubernetes.io/projected/1ce971f4-45ea-463f-82da-687f425ddfb6-kube-api-access-dtzt9\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.160930 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-utilities\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.161440 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-utilities\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.162113 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-catalog-content\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.186475 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtzt9\" (UniqueName: \"kubernetes.io/projected/1ce971f4-45ea-463f-82da-687f425ddfb6-kube-api-access-dtzt9\") pod \"certified-operators-scpvw\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.256046 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:28 crc kubenswrapper[4935]: I1005 08:23:28.762461 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-scpvw"] Oct 05 08:23:29 crc kubenswrapper[4935]: I1005 08:23:29.279146 4935 generic.go:334] "Generic (PLEG): container finished" podID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerID="6d9c73df8d905ce95c91fc2c98070637009ce2c57595df8b08a5bee345036ff7" exitCode=0 Oct 05 08:23:29 crc kubenswrapper[4935]: I1005 08:23:29.279314 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-scpvw" event={"ID":"1ce971f4-45ea-463f-82da-687f425ddfb6","Type":"ContainerDied","Data":"6d9c73df8d905ce95c91fc2c98070637009ce2c57595df8b08a5bee345036ff7"} Oct 05 08:23:29 crc kubenswrapper[4935]: I1005 08:23:29.279794 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-scpvw" event={"ID":"1ce971f4-45ea-463f-82da-687f425ddfb6","Type":"ContainerStarted","Data":"b39be050e9842a2edf23891bffaddd4c9239ef8d3e28528c80097fe63df221aa"} Oct 05 08:23:30 crc kubenswrapper[4935]: I1005 08:23:30.292243 4935 generic.go:334] "Generic (PLEG): container finished" podID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerID="65b2d31292198f40b2c2e30e1dbd7e9f99d2a285eab012deb88080f139332ef5" exitCode=0 Oct 05 08:23:30 crc kubenswrapper[4935]: I1005 08:23:30.292299 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-scpvw" event={"ID":"1ce971f4-45ea-463f-82da-687f425ddfb6","Type":"ContainerDied","Data":"65b2d31292198f40b2c2e30e1dbd7e9f99d2a285eab012deb88080f139332ef5"} Oct 05 08:23:31 crc kubenswrapper[4935]: I1005 08:23:31.302046 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-scpvw" event={"ID":"1ce971f4-45ea-463f-82da-687f425ddfb6","Type":"ContainerStarted","Data":"5991b2603a6ffdff2781f0b639a437618a2f93cd5ca7e6288b1658d4f4864089"} Oct 05 08:23:31 crc kubenswrapper[4935]: I1005 08:23:31.334101 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-scpvw" podStartSLOduration=2.944952206 podStartE2EDuration="4.334078129s" podCreationTimestamp="2025-10-05 08:23:27 +0000 UTC" firstStartedPulling="2025-10-05 08:23:29.282129513 +0000 UTC m=+5443.164755983" lastFinishedPulling="2025-10-05 08:23:30.671255456 +0000 UTC m=+5444.553881906" observedRunningTime="2025-10-05 08:23:31.326139679 +0000 UTC m=+5445.208766149" watchObservedRunningTime="2025-10-05 08:23:31.334078129 +0000 UTC m=+5445.216704589" Oct 05 08:23:38 crc kubenswrapper[4935]: I1005 08:23:38.256597 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:38 crc kubenswrapper[4935]: I1005 08:23:38.257223 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:38 crc kubenswrapper[4935]: I1005 08:23:38.319089 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:38 crc kubenswrapper[4935]: I1005 08:23:38.421298 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.056997 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kjgc2"] Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.058814 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.081612 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kjgc2"] Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.237717 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-catalog-content\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.237848 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84btx\" (UniqueName: \"kubernetes.io/projected/eec88366-eb48-4277-83a2-5a1508558abb-kube-api-access-84btx\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.238015 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-utilities\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.339038 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-catalog-content\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.339112 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84btx\" (UniqueName: \"kubernetes.io/projected/eec88366-eb48-4277-83a2-5a1508558abb-kube-api-access-84btx\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.339166 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-utilities\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.339573 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-catalog-content\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.339619 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-utilities\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.374706 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84btx\" (UniqueName: \"kubernetes.io/projected/eec88366-eb48-4277-83a2-5a1508558abb-kube-api-access-84btx\") pod \"community-operators-kjgc2\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.394444 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:40 crc kubenswrapper[4935]: I1005 08:23:40.944514 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kjgc2"] Oct 05 08:23:40 crc kubenswrapper[4935]: W1005 08:23:40.952550 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeec88366_eb48_4277_83a2_5a1508558abb.slice/crio-cc0269dfd1a5c74630ba5840af33ea3b1514db01eef34be5a5a91757f273536d WatchSource:0}: Error finding container cc0269dfd1a5c74630ba5840af33ea3b1514db01eef34be5a5a91757f273536d: Status 404 returned error can't find the container with id cc0269dfd1a5c74630ba5840af33ea3b1514db01eef34be5a5a91757f273536d Oct 05 08:23:41 crc kubenswrapper[4935]: I1005 08:23:41.393687 4935 generic.go:334] "Generic (PLEG): container finished" podID="eec88366-eb48-4277-83a2-5a1508558abb" containerID="289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975" exitCode=0 Oct 05 08:23:41 crc kubenswrapper[4935]: I1005 08:23:41.393828 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjgc2" event={"ID":"eec88366-eb48-4277-83a2-5a1508558abb","Type":"ContainerDied","Data":"289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975"} Oct 05 08:23:41 crc kubenswrapper[4935]: I1005 08:23:41.394283 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjgc2" event={"ID":"eec88366-eb48-4277-83a2-5a1508558abb","Type":"ContainerStarted","Data":"cc0269dfd1a5c74630ba5840af33ea3b1514db01eef34be5a5a91757f273536d"} Oct 05 08:23:43 crc kubenswrapper[4935]: I1005 08:23:43.421402 4935 generic.go:334] "Generic (PLEG): container finished" podID="eec88366-eb48-4277-83a2-5a1508558abb" containerID="5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768" exitCode=0 Oct 05 08:23:43 crc kubenswrapper[4935]: I1005 08:23:43.421558 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjgc2" event={"ID":"eec88366-eb48-4277-83a2-5a1508558abb","Type":"ContainerDied","Data":"5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768"} Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.252899 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-scpvw"] Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.253424 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-scpvw" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerName="registry-server" containerID="cri-o://5991b2603a6ffdff2781f0b639a437618a2f93cd5ca7e6288b1658d4f4864089" gracePeriod=2 Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.430280 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjgc2" event={"ID":"eec88366-eb48-4277-83a2-5a1508558abb","Type":"ContainerStarted","Data":"192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3"} Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.432719 4935 generic.go:334] "Generic (PLEG): container finished" podID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerID="5991b2603a6ffdff2781f0b639a437618a2f93cd5ca7e6288b1658d4f4864089" exitCode=0 Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.432761 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-scpvw" event={"ID":"1ce971f4-45ea-463f-82da-687f425ddfb6","Type":"ContainerDied","Data":"5991b2603a6ffdff2781f0b639a437618a2f93cd5ca7e6288b1658d4f4864089"} Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.448813 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kjgc2" podStartSLOduration=1.765388889 podStartE2EDuration="4.448795549s" podCreationTimestamp="2025-10-05 08:23:40 +0000 UTC" firstStartedPulling="2025-10-05 08:23:41.397028515 +0000 UTC m=+5455.279655005" lastFinishedPulling="2025-10-05 08:23:44.080435165 +0000 UTC m=+5457.963061665" observedRunningTime="2025-10-05 08:23:44.448131241 +0000 UTC m=+5458.330757701" watchObservedRunningTime="2025-10-05 08:23:44.448795549 +0000 UTC m=+5458.331422009" Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.695843 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.805825 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtzt9\" (UniqueName: \"kubernetes.io/projected/1ce971f4-45ea-463f-82da-687f425ddfb6-kube-api-access-dtzt9\") pod \"1ce971f4-45ea-463f-82da-687f425ddfb6\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.806014 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-catalog-content\") pod \"1ce971f4-45ea-463f-82da-687f425ddfb6\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.806133 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-utilities\") pod \"1ce971f4-45ea-463f-82da-687f425ddfb6\" (UID: \"1ce971f4-45ea-463f-82da-687f425ddfb6\") " Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.807065 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-utilities" (OuterVolumeSpecName: "utilities") pod "1ce971f4-45ea-463f-82da-687f425ddfb6" (UID: "1ce971f4-45ea-463f-82da-687f425ddfb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.814923 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ce971f4-45ea-463f-82da-687f425ddfb6-kube-api-access-dtzt9" (OuterVolumeSpecName: "kube-api-access-dtzt9") pod "1ce971f4-45ea-463f-82da-687f425ddfb6" (UID: "1ce971f4-45ea-463f-82da-687f425ddfb6"). InnerVolumeSpecName "kube-api-access-dtzt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.853138 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ce971f4-45ea-463f-82da-687f425ddfb6" (UID: "1ce971f4-45ea-463f-82da-687f425ddfb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.908156 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtzt9\" (UniqueName: \"kubernetes.io/projected/1ce971f4-45ea-463f-82da-687f425ddfb6-kube-api-access-dtzt9\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.908197 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:44 crc kubenswrapper[4935]: I1005 08:23:44.908208 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce971f4-45ea-463f-82da-687f425ddfb6-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:45 crc kubenswrapper[4935]: I1005 08:23:45.444634 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-scpvw" Oct 05 08:23:45 crc kubenswrapper[4935]: I1005 08:23:45.444686 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-scpvw" event={"ID":"1ce971f4-45ea-463f-82da-687f425ddfb6","Type":"ContainerDied","Data":"b39be050e9842a2edf23891bffaddd4c9239ef8d3e28528c80097fe63df221aa"} Oct 05 08:23:45 crc kubenswrapper[4935]: I1005 08:23:45.444747 4935 scope.go:117] "RemoveContainer" containerID="5991b2603a6ffdff2781f0b639a437618a2f93cd5ca7e6288b1658d4f4864089" Oct 05 08:23:45 crc kubenswrapper[4935]: I1005 08:23:45.472014 4935 scope.go:117] "RemoveContainer" containerID="65b2d31292198f40b2c2e30e1dbd7e9f99d2a285eab012deb88080f139332ef5" Oct 05 08:23:45 crc kubenswrapper[4935]: I1005 08:23:45.489114 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-scpvw"] Oct 05 08:23:45 crc kubenswrapper[4935]: I1005 08:23:45.493208 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-scpvw"] Oct 05 08:23:45 crc kubenswrapper[4935]: I1005 08:23:45.520685 4935 scope.go:117] "RemoveContainer" containerID="6d9c73df8d905ce95c91fc2c98070637009ce2c57595df8b08a5bee345036ff7" Oct 05 08:23:46 crc kubenswrapper[4935]: I1005 08:23:46.787683 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" path="/var/lib/kubelet/pods/1ce971f4-45ea-463f-82da-687f425ddfb6/volumes" Oct 05 08:23:50 crc kubenswrapper[4935]: I1005 08:23:50.394659 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:50 crc kubenswrapper[4935]: I1005 08:23:50.395250 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:50 crc kubenswrapper[4935]: I1005 08:23:50.463173 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:50 crc kubenswrapper[4935]: I1005 08:23:50.513430 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:52 crc kubenswrapper[4935]: I1005 08:23:52.258056 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kjgc2"] Oct 05 08:23:52 crc kubenswrapper[4935]: I1005 08:23:52.494839 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kjgc2" podUID="eec88366-eb48-4277-83a2-5a1508558abb" containerName="registry-server" containerID="cri-o://192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3" gracePeriod=2 Oct 05 08:23:52 crc kubenswrapper[4935]: I1005 08:23:52.943450 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.026829 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84btx\" (UniqueName: \"kubernetes.io/projected/eec88366-eb48-4277-83a2-5a1508558abb-kube-api-access-84btx\") pod \"eec88366-eb48-4277-83a2-5a1508558abb\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.026880 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-catalog-content\") pod \"eec88366-eb48-4277-83a2-5a1508558abb\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.026966 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-utilities\") pod \"eec88366-eb48-4277-83a2-5a1508558abb\" (UID: \"eec88366-eb48-4277-83a2-5a1508558abb\") " Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.028204 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-utilities" (OuterVolumeSpecName: "utilities") pod "eec88366-eb48-4277-83a2-5a1508558abb" (UID: "eec88366-eb48-4277-83a2-5a1508558abb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.032124 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eec88366-eb48-4277-83a2-5a1508558abb-kube-api-access-84btx" (OuterVolumeSpecName: "kube-api-access-84btx") pod "eec88366-eb48-4277-83a2-5a1508558abb" (UID: "eec88366-eb48-4277-83a2-5a1508558abb"). InnerVolumeSpecName "kube-api-access-84btx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.107459 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eec88366-eb48-4277-83a2-5a1508558abb" (UID: "eec88366-eb48-4277-83a2-5a1508558abb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.128769 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84btx\" (UniqueName: \"kubernetes.io/projected/eec88366-eb48-4277-83a2-5a1508558abb-kube-api-access-84btx\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.128830 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.128844 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec88366-eb48-4277-83a2-5a1508558abb-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.503322 4935 generic.go:334] "Generic (PLEG): container finished" podID="eec88366-eb48-4277-83a2-5a1508558abb" containerID="192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3" exitCode=0 Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.503373 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjgc2" event={"ID":"eec88366-eb48-4277-83a2-5a1508558abb","Type":"ContainerDied","Data":"192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3"} Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.503400 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjgc2" event={"ID":"eec88366-eb48-4277-83a2-5a1508558abb","Type":"ContainerDied","Data":"cc0269dfd1a5c74630ba5840af33ea3b1514db01eef34be5a5a91757f273536d"} Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.503426 4935 scope.go:117] "RemoveContainer" containerID="192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.503457 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjgc2" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.525195 4935 scope.go:117] "RemoveContainer" containerID="5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.550786 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kjgc2"] Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.558443 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kjgc2"] Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.571859 4935 scope.go:117] "RemoveContainer" containerID="289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.602573 4935 scope.go:117] "RemoveContainer" containerID="192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3" Oct 05 08:23:53 crc kubenswrapper[4935]: E1005 08:23:53.603214 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3\": container with ID starting with 192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3 not found: ID does not exist" containerID="192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.603267 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3"} err="failed to get container status \"192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3\": rpc error: code = NotFound desc = could not find container \"192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3\": container with ID starting with 192d977c2d5416b3293a34a14ced78a833afdd1459bd9fba8bab273dd59914c3 not found: ID does not exist" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.603299 4935 scope.go:117] "RemoveContainer" containerID="5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768" Oct 05 08:23:53 crc kubenswrapper[4935]: E1005 08:23:53.603675 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768\": container with ID starting with 5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768 not found: ID does not exist" containerID="5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.603711 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768"} err="failed to get container status \"5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768\": rpc error: code = NotFound desc = could not find container \"5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768\": container with ID starting with 5dbe403f02b84aa77d8117ae07e0229d53feab9b864c43e1b021dd0c733d7768 not found: ID does not exist" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.603734 4935 scope.go:117] "RemoveContainer" containerID="289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975" Oct 05 08:23:53 crc kubenswrapper[4935]: E1005 08:23:53.604187 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975\": container with ID starting with 289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975 not found: ID does not exist" containerID="289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975" Oct 05 08:23:53 crc kubenswrapper[4935]: I1005 08:23:53.604258 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975"} err="failed to get container status \"289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975\": rpc error: code = NotFound desc = could not find container \"289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975\": container with ID starting with 289e20a9dba4cfd0c20a807c872e674ac1a2dcbb800ea2e3756493aa4d634975 not found: ID does not exist" Oct 05 08:23:54 crc kubenswrapper[4935]: I1005 08:23:54.786726 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eec88366-eb48-4277-83a2-5a1508558abb" path="/var/lib/kubelet/pods/eec88366-eb48-4277-83a2-5a1508558abb/volumes" Oct 05 08:25:14 crc kubenswrapper[4935]: I1005 08:25:14.289222 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:25:14 crc kubenswrapper[4935]: I1005 08:25:14.289864 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:25:44 crc kubenswrapper[4935]: I1005 08:25:44.289442 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:25:44 crc kubenswrapper[4935]: I1005 08:25:44.290073 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.289969 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.291035 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.291127 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.292252 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.292502 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" gracePeriod=600 Oct 05 08:26:14 crc kubenswrapper[4935]: E1005 08:26:14.426379 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.834989 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" exitCode=0 Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.835103 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484"} Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.835387 4935 scope.go:117] "RemoveContainer" containerID="ed3a0b66dfca5c0c324f5e00aad5979dd80a4661554edeae46ce1aa1c45c00be" Oct 05 08:26:14 crc kubenswrapper[4935]: I1005 08:26:14.837316 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:26:14 crc kubenswrapper[4935]: E1005 08:26:14.837736 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:26:26 crc kubenswrapper[4935]: I1005 08:26:26.785341 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:26:26 crc kubenswrapper[4935]: E1005 08:26:26.786111 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.267399 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n2j4l"] Oct 05 08:26:33 crc kubenswrapper[4935]: E1005 08:26:33.269082 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec88366-eb48-4277-83a2-5a1508558abb" containerName="extract-content" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.269122 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec88366-eb48-4277-83a2-5a1508558abb" containerName="extract-content" Oct 05 08:26:33 crc kubenswrapper[4935]: E1005 08:26:33.269176 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerName="extract-utilities" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.269194 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerName="extract-utilities" Oct 05 08:26:33 crc kubenswrapper[4935]: E1005 08:26:33.269256 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerName="extract-content" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.269277 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerName="extract-content" Oct 05 08:26:33 crc kubenswrapper[4935]: E1005 08:26:33.269300 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec88366-eb48-4277-83a2-5a1508558abb" containerName="registry-server" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.269317 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec88366-eb48-4277-83a2-5a1508558abb" containerName="registry-server" Oct 05 08:26:33 crc kubenswrapper[4935]: E1005 08:26:33.269344 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec88366-eb48-4277-83a2-5a1508558abb" containerName="extract-utilities" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.269362 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec88366-eb48-4277-83a2-5a1508558abb" containerName="extract-utilities" Oct 05 08:26:33 crc kubenswrapper[4935]: E1005 08:26:33.269389 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerName="registry-server" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.269406 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerName="registry-server" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.269990 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ce971f4-45ea-463f-82da-687f425ddfb6" containerName="registry-server" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.270030 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec88366-eb48-4277-83a2-5a1508558abb" containerName="registry-server" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.272855 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.273416 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2j4l"] Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.304842 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-utilities\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.304973 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzmdz\" (UniqueName: \"kubernetes.io/projected/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-kube-api-access-mzmdz\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.305119 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-catalog-content\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.405816 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzmdz\" (UniqueName: \"kubernetes.io/projected/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-kube-api-access-mzmdz\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.405928 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-catalog-content\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.405957 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-utilities\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.406407 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-utilities\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.406682 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-catalog-content\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.436148 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzmdz\" (UniqueName: \"kubernetes.io/projected/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-kube-api-access-mzmdz\") pod \"redhat-marketplace-n2j4l\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:33 crc kubenswrapper[4935]: I1005 08:26:33.601620 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:34 crc kubenswrapper[4935]: I1005 08:26:34.058341 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2j4l"] Oct 05 08:26:35 crc kubenswrapper[4935]: I1005 08:26:35.028287 4935 generic.go:334] "Generic (PLEG): container finished" podID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerID="c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24" exitCode=0 Oct 05 08:26:35 crc kubenswrapper[4935]: I1005 08:26:35.028393 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2j4l" event={"ID":"cbde9a6e-f695-45ac-8925-cdaebf03ab0e","Type":"ContainerDied","Data":"c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24"} Oct 05 08:26:35 crc kubenswrapper[4935]: I1005 08:26:35.028618 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2j4l" event={"ID":"cbde9a6e-f695-45ac-8925-cdaebf03ab0e","Type":"ContainerStarted","Data":"fd4b4db44441aac295dfd663a22c52a9a77e073107eddeb082022dbaf451753c"} Oct 05 08:26:35 crc kubenswrapper[4935]: I1005 08:26:35.031388 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:26:36 crc kubenswrapper[4935]: I1005 08:26:36.040712 4935 generic.go:334] "Generic (PLEG): container finished" podID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerID="99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91" exitCode=0 Oct 05 08:26:36 crc kubenswrapper[4935]: I1005 08:26:36.040811 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2j4l" event={"ID":"cbde9a6e-f695-45ac-8925-cdaebf03ab0e","Type":"ContainerDied","Data":"99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91"} Oct 05 08:26:37 crc kubenswrapper[4935]: I1005 08:26:37.059235 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2j4l" event={"ID":"cbde9a6e-f695-45ac-8925-cdaebf03ab0e","Type":"ContainerStarted","Data":"4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845"} Oct 05 08:26:37 crc kubenswrapper[4935]: I1005 08:26:37.085538 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n2j4l" podStartSLOduration=2.46503612 podStartE2EDuration="4.085513932s" podCreationTimestamp="2025-10-05 08:26:33 +0000 UTC" firstStartedPulling="2025-10-05 08:26:35.030773863 +0000 UTC m=+5628.913400353" lastFinishedPulling="2025-10-05 08:26:36.651251705 +0000 UTC m=+5630.533878165" observedRunningTime="2025-10-05 08:26:37.080326715 +0000 UTC m=+5630.962953185" watchObservedRunningTime="2025-10-05 08:26:37.085513932 +0000 UTC m=+5630.968140392" Oct 05 08:26:38 crc kubenswrapper[4935]: I1005 08:26:38.776957 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:26:38 crc kubenswrapper[4935]: E1005 08:26:38.777172 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:26:43 crc kubenswrapper[4935]: I1005 08:26:43.601980 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:43 crc kubenswrapper[4935]: I1005 08:26:43.602387 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:43 crc kubenswrapper[4935]: I1005 08:26:43.659009 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:44 crc kubenswrapper[4935]: I1005 08:26:44.161708 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:44 crc kubenswrapper[4935]: I1005 08:26:44.208136 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2j4l"] Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.140183 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n2j4l" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerName="registry-server" containerID="cri-o://4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845" gracePeriod=2 Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.613436 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.711859 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzmdz\" (UniqueName: \"kubernetes.io/projected/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-kube-api-access-mzmdz\") pod \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.711967 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-catalog-content\") pod \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.712046 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-utilities\") pod \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\" (UID: \"cbde9a6e-f695-45ac-8925-cdaebf03ab0e\") " Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.713816 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-utilities" (OuterVolumeSpecName: "utilities") pod "cbde9a6e-f695-45ac-8925-cdaebf03ab0e" (UID: "cbde9a6e-f695-45ac-8925-cdaebf03ab0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.722071 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-kube-api-access-mzmdz" (OuterVolumeSpecName: "kube-api-access-mzmdz") pod "cbde9a6e-f695-45ac-8925-cdaebf03ab0e" (UID: "cbde9a6e-f695-45ac-8925-cdaebf03ab0e"). InnerVolumeSpecName "kube-api-access-mzmdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.733730 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cbde9a6e-f695-45ac-8925-cdaebf03ab0e" (UID: "cbde9a6e-f695-45ac-8925-cdaebf03ab0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.813849 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.813954 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzmdz\" (UniqueName: \"kubernetes.io/projected/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-kube-api-access-mzmdz\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:46 crc kubenswrapper[4935]: I1005 08:26:46.813977 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbde9a6e-f695-45ac-8925-cdaebf03ab0e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.159271 4935 generic.go:334] "Generic (PLEG): container finished" podID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerID="4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845" exitCode=0 Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.159350 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2j4l" event={"ID":"cbde9a6e-f695-45ac-8925-cdaebf03ab0e","Type":"ContainerDied","Data":"4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845"} Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.159411 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2j4l" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.159450 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2j4l" event={"ID":"cbde9a6e-f695-45ac-8925-cdaebf03ab0e","Type":"ContainerDied","Data":"fd4b4db44441aac295dfd663a22c52a9a77e073107eddeb082022dbaf451753c"} Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.159508 4935 scope.go:117] "RemoveContainer" containerID="4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.190429 4935 scope.go:117] "RemoveContainer" containerID="99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.196005 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2j4l"] Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.202363 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2j4l"] Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.208253 4935 scope.go:117] "RemoveContainer" containerID="c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.241981 4935 scope.go:117] "RemoveContainer" containerID="4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845" Oct 05 08:26:47 crc kubenswrapper[4935]: E1005 08:26:47.242528 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845\": container with ID starting with 4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845 not found: ID does not exist" containerID="4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.242573 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845"} err="failed to get container status \"4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845\": rpc error: code = NotFound desc = could not find container \"4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845\": container with ID starting with 4d4631260a563b38dccdfb1fdb04e297a793fff50e67aea820317858f60a2845 not found: ID does not exist" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.242598 4935 scope.go:117] "RemoveContainer" containerID="99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91" Oct 05 08:26:47 crc kubenswrapper[4935]: E1005 08:26:47.243022 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91\": container with ID starting with 99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91 not found: ID does not exist" containerID="99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.243047 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91"} err="failed to get container status \"99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91\": rpc error: code = NotFound desc = could not find container \"99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91\": container with ID starting with 99159d561d40586f764aebd8c44b22993cceef6d31a1c170cf82ac884b8b5e91 not found: ID does not exist" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.243062 4935 scope.go:117] "RemoveContainer" containerID="c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24" Oct 05 08:26:47 crc kubenswrapper[4935]: E1005 08:26:47.243328 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24\": container with ID starting with c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24 not found: ID does not exist" containerID="c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24" Oct 05 08:26:47 crc kubenswrapper[4935]: I1005 08:26:47.243358 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24"} err="failed to get container status \"c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24\": rpc error: code = NotFound desc = could not find container \"c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24\": container with ID starting with c16429d837d8b0ccd6abc0f0beaa2f1bccfa2fedcdd7ac92d5b924a150eece24 not found: ID does not exist" Oct 05 08:26:48 crc kubenswrapper[4935]: I1005 08:26:48.786339 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" path="/var/lib/kubelet/pods/cbde9a6e-f695-45ac-8925-cdaebf03ab0e/volumes" Oct 05 08:26:53 crc kubenswrapper[4935]: I1005 08:26:53.777835 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:26:53 crc kubenswrapper[4935]: E1005 08:26:53.779688 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:27:07 crc kubenswrapper[4935]: I1005 08:27:07.777864 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:27:07 crc kubenswrapper[4935]: E1005 08:27:07.779465 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:27:19 crc kubenswrapper[4935]: I1005 08:27:19.777137 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:27:19 crc kubenswrapper[4935]: E1005 08:27:19.779561 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:27:34 crc kubenswrapper[4935]: I1005 08:27:34.777299 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:27:34 crc kubenswrapper[4935]: E1005 08:27:34.777803 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:27:49 crc kubenswrapper[4935]: I1005 08:27:49.777391 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:27:49 crc kubenswrapper[4935]: E1005 08:27:49.778213 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:28:00 crc kubenswrapper[4935]: I1005 08:28:00.777255 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:28:00 crc kubenswrapper[4935]: E1005 08:28:00.778196 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:28:12 crc kubenswrapper[4935]: I1005 08:28:12.777604 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:28:12 crc kubenswrapper[4935]: E1005 08:28:12.778581 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:28:25 crc kubenswrapper[4935]: I1005 08:28:25.776778 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:28:25 crc kubenswrapper[4935]: E1005 08:28:25.777550 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:28:36 crc kubenswrapper[4935]: I1005 08:28:36.791860 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:28:36 crc kubenswrapper[4935]: E1005 08:28:36.793859 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.731008 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-j2hvq"] Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.737879 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-j2hvq"] Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.788116 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a62d002c-6756-491b-8935-02235d870828" path="/var/lib/kubelet/pods/a62d002c-6756-491b-8935-02235d870828/volumes" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.860578 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-jgd27"] Oct 05 08:28:38 crc kubenswrapper[4935]: E1005 08:28:38.861068 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerName="extract-content" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.861106 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerName="extract-content" Oct 05 08:28:38 crc kubenswrapper[4935]: E1005 08:28:38.861168 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerName="extract-utilities" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.861188 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerName="extract-utilities" Oct 05 08:28:38 crc kubenswrapper[4935]: E1005 08:28:38.861222 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerName="registry-server" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.861240 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerName="registry-server" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.861554 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbde9a6e-f695-45ac-8925-cdaebf03ab0e" containerName="registry-server" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.862350 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.864014 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.866188 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.866360 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.866430 4935 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-qpj25" Oct 05 08:28:38 crc kubenswrapper[4935]: I1005 08:28:38.881458 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-jgd27"] Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.028717 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkxdk\" (UniqueName: \"kubernetes.io/projected/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-kube-api-access-dkxdk\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.028784 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-crc-storage\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.028807 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-node-mnt\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.130003 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkxdk\" (UniqueName: \"kubernetes.io/projected/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-kube-api-access-dkxdk\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.130070 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-crc-storage\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.130106 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-node-mnt\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.130497 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-node-mnt\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.131069 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-crc-storage\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.155690 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkxdk\" (UniqueName: \"kubernetes.io/projected/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-kube-api-access-dkxdk\") pod \"crc-storage-crc-jgd27\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.179458 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:39 crc kubenswrapper[4935]: I1005 08:28:39.651991 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-jgd27"] Oct 05 08:28:40 crc kubenswrapper[4935]: I1005 08:28:40.193854 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-jgd27" event={"ID":"13a8e38c-d51c-46cf-a0af-4b662a3e3f34","Type":"ContainerStarted","Data":"f4e406fb75dddc30f08c304da3edc120e22e3a2c542d5ad56df01c2e6d99eef8"} Oct 05 08:28:41 crc kubenswrapper[4935]: I1005 08:28:41.203540 4935 generic.go:334] "Generic (PLEG): container finished" podID="13a8e38c-d51c-46cf-a0af-4b662a3e3f34" containerID="8af944ec27a78ab87289bc3295023b1f25f79ee198bedd758226ec4a7b425ef8" exitCode=0 Oct 05 08:28:41 crc kubenswrapper[4935]: I1005 08:28:41.203610 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-jgd27" event={"ID":"13a8e38c-d51c-46cf-a0af-4b662a3e3f34","Type":"ContainerDied","Data":"8af944ec27a78ab87289bc3295023b1f25f79ee198bedd758226ec4a7b425ef8"} Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.584183 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.682678 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-crc-storage\") pod \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.682741 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-node-mnt\") pod \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.682761 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkxdk\" (UniqueName: \"kubernetes.io/projected/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-kube-api-access-dkxdk\") pod \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\" (UID: \"13a8e38c-d51c-46cf-a0af-4b662a3e3f34\") " Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.682885 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "13a8e38c-d51c-46cf-a0af-4b662a3e3f34" (UID: "13a8e38c-d51c-46cf-a0af-4b662a3e3f34"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.690157 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-kube-api-access-dkxdk" (OuterVolumeSpecName: "kube-api-access-dkxdk") pod "13a8e38c-d51c-46cf-a0af-4b662a3e3f34" (UID: "13a8e38c-d51c-46cf-a0af-4b662a3e3f34"). InnerVolumeSpecName "kube-api-access-dkxdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.727640 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "13a8e38c-d51c-46cf-a0af-4b662a3e3f34" (UID: "13a8e38c-d51c-46cf-a0af-4b662a3e3f34"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.783869 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkxdk\" (UniqueName: \"kubernetes.io/projected/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-kube-api-access-dkxdk\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.783907 4935 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:42.783918 4935 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/13a8e38c-d51c-46cf-a0af-4b662a3e3f34-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:43.227190 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-jgd27" event={"ID":"13a8e38c-d51c-46cf-a0af-4b662a3e3f34","Type":"ContainerDied","Data":"f4e406fb75dddc30f08c304da3edc120e22e3a2c542d5ad56df01c2e6d99eef8"} Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:43.227261 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4e406fb75dddc30f08c304da3edc120e22e3a2c542d5ad56df01c2e6d99eef8" Oct 05 08:28:43 crc kubenswrapper[4935]: I1005 08:28:43.227266 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-jgd27" Oct 05 08:28:44 crc kubenswrapper[4935]: I1005 08:28:44.873951 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-jgd27"] Oct 05 08:28:44 crc kubenswrapper[4935]: I1005 08:28:44.885081 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-jgd27"] Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.070637 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-7qvtn"] Oct 05 08:28:45 crc kubenswrapper[4935]: E1005 08:28:45.071178 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13a8e38c-d51c-46cf-a0af-4b662a3e3f34" containerName="storage" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.071206 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="13a8e38c-d51c-46cf-a0af-4b662a3e3f34" containerName="storage" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.071392 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="13a8e38c-d51c-46cf-a0af-4b662a3e3f34" containerName="storage" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.072188 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.074739 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.076797 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.079720 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.080176 4935 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-qpj25" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.081466 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-7qvtn"] Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.216879 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0775f30c-4957-4c85-b31b-f87581d89a33-crc-storage\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.216996 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89fw5\" (UniqueName: \"kubernetes.io/projected/0775f30c-4957-4c85-b31b-f87581d89a33-kube-api-access-89fw5\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.217526 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0775f30c-4957-4c85-b31b-f87581d89a33-node-mnt\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.318774 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0775f30c-4957-4c85-b31b-f87581d89a33-crc-storage\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.318915 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89fw5\" (UniqueName: \"kubernetes.io/projected/0775f30c-4957-4c85-b31b-f87581d89a33-kube-api-access-89fw5\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.318960 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0775f30c-4957-4c85-b31b-f87581d89a33-node-mnt\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.319320 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0775f30c-4957-4c85-b31b-f87581d89a33-node-mnt\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.319730 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0775f30c-4957-4c85-b31b-f87581d89a33-crc-storage\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.345151 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89fw5\" (UniqueName: \"kubernetes.io/projected/0775f30c-4957-4c85-b31b-f87581d89a33-kube-api-access-89fw5\") pod \"crc-storage-crc-7qvtn\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.405094 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:45 crc kubenswrapper[4935]: I1005 08:28:45.811875 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-7qvtn"] Oct 05 08:28:46 crc kubenswrapper[4935]: I1005 08:28:46.255823 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7qvtn" event={"ID":"0775f30c-4957-4c85-b31b-f87581d89a33","Type":"ContainerStarted","Data":"277bd44146273f065d09b31efca0d48dd5d77431f164e68f3f24c2ad46a9b12f"} Oct 05 08:28:46 crc kubenswrapper[4935]: I1005 08:28:46.792948 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13a8e38c-d51c-46cf-a0af-4b662a3e3f34" path="/var/lib/kubelet/pods/13a8e38c-d51c-46cf-a0af-4b662a3e3f34/volumes" Oct 05 08:28:47 crc kubenswrapper[4935]: I1005 08:28:47.265199 4935 generic.go:334] "Generic (PLEG): container finished" podID="0775f30c-4957-4c85-b31b-f87581d89a33" containerID="ca0ba3770ee2057eb0741918bde3e860bd922b5604e481e98820f111d6a329d7" exitCode=0 Oct 05 08:28:47 crc kubenswrapper[4935]: I1005 08:28:47.265242 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7qvtn" event={"ID":"0775f30c-4957-4c85-b31b-f87581d89a33","Type":"ContainerDied","Data":"ca0ba3770ee2057eb0741918bde3e860bd922b5604e481e98820f111d6a329d7"} Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.556749 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.566002 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0775f30c-4957-4c85-b31b-f87581d89a33-crc-storage\") pod \"0775f30c-4957-4c85-b31b-f87581d89a33\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.566120 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89fw5\" (UniqueName: \"kubernetes.io/projected/0775f30c-4957-4c85-b31b-f87581d89a33-kube-api-access-89fw5\") pod \"0775f30c-4957-4c85-b31b-f87581d89a33\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.566164 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0775f30c-4957-4c85-b31b-f87581d89a33-node-mnt\") pod \"0775f30c-4957-4c85-b31b-f87581d89a33\" (UID: \"0775f30c-4957-4c85-b31b-f87581d89a33\") " Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.566602 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0775f30c-4957-4c85-b31b-f87581d89a33-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "0775f30c-4957-4c85-b31b-f87581d89a33" (UID: "0775f30c-4957-4c85-b31b-f87581d89a33"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.576141 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0775f30c-4957-4c85-b31b-f87581d89a33-kube-api-access-89fw5" (OuterVolumeSpecName: "kube-api-access-89fw5") pod "0775f30c-4957-4c85-b31b-f87581d89a33" (UID: "0775f30c-4957-4c85-b31b-f87581d89a33"). InnerVolumeSpecName "kube-api-access-89fw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.589540 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0775f30c-4957-4c85-b31b-f87581d89a33-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "0775f30c-4957-4c85-b31b-f87581d89a33" (UID: "0775f30c-4957-4c85-b31b-f87581d89a33"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.667738 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89fw5\" (UniqueName: \"kubernetes.io/projected/0775f30c-4957-4c85-b31b-f87581d89a33-kube-api-access-89fw5\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.667777 4935 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0775f30c-4957-4c85-b31b-f87581d89a33-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:48 crc kubenswrapper[4935]: I1005 08:28:48.667788 4935 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0775f30c-4957-4c85-b31b-f87581d89a33-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:49 crc kubenswrapper[4935]: I1005 08:28:49.280615 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7qvtn" event={"ID":"0775f30c-4957-4c85-b31b-f87581d89a33","Type":"ContainerDied","Data":"277bd44146273f065d09b31efca0d48dd5d77431f164e68f3f24c2ad46a9b12f"} Oct 05 08:28:49 crc kubenswrapper[4935]: I1005 08:28:49.280659 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="277bd44146273f065d09b31efca0d48dd5d77431f164e68f3f24c2ad46a9b12f" Oct 05 08:28:49 crc kubenswrapper[4935]: I1005 08:28:49.280661 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7qvtn" Oct 05 08:28:50 crc kubenswrapper[4935]: I1005 08:28:50.776787 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:28:50 crc kubenswrapper[4935]: E1005 08:28:50.777332 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:28:56 crc kubenswrapper[4935]: I1005 08:28:56.289972 4935 scope.go:117] "RemoveContainer" containerID="d7ffcb2306d39f5d2b1224f80f8160485b1405feabc0a2bd0fb06694d9f3af3a" Oct 05 08:29:02 crc kubenswrapper[4935]: I1005 08:29:02.778218 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:29:02 crc kubenswrapper[4935]: E1005 08:29:02.779544 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:29:14 crc kubenswrapper[4935]: I1005 08:29:14.777862 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:29:14 crc kubenswrapper[4935]: E1005 08:29:14.778860 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:29:27 crc kubenswrapper[4935]: I1005 08:29:27.778021 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:29:27 crc kubenswrapper[4935]: E1005 08:29:27.779215 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:29:40 crc kubenswrapper[4935]: I1005 08:29:40.778360 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:29:40 crc kubenswrapper[4935]: E1005 08:29:40.779029 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:29:53 crc kubenswrapper[4935]: I1005 08:29:53.777767 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:29:53 crc kubenswrapper[4935]: E1005 08:29:53.778876 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.161033 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95"] Oct 05 08:30:00 crc kubenswrapper[4935]: E1005 08:30:00.164534 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0775f30c-4957-4c85-b31b-f87581d89a33" containerName="storage" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.164576 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0775f30c-4957-4c85-b31b-f87581d89a33" containerName="storage" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.164812 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0775f30c-4957-4c85-b31b-f87581d89a33" containerName="storage" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.165595 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.168638 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.172858 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.185045 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95"] Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.243426 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlbg7\" (UniqueName: \"kubernetes.io/projected/99619e61-ede2-432a-a293-5fab27d7d60c-kube-api-access-tlbg7\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.243614 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99619e61-ede2-432a-a293-5fab27d7d60c-config-volume\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.243665 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99619e61-ede2-432a-a293-5fab27d7d60c-secret-volume\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.345660 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99619e61-ede2-432a-a293-5fab27d7d60c-config-volume\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.345740 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99619e61-ede2-432a-a293-5fab27d7d60c-secret-volume\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.345817 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlbg7\" (UniqueName: \"kubernetes.io/projected/99619e61-ede2-432a-a293-5fab27d7d60c-kube-api-access-tlbg7\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.347418 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99619e61-ede2-432a-a293-5fab27d7d60c-config-volume\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.360090 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99619e61-ede2-432a-a293-5fab27d7d60c-secret-volume\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.366594 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlbg7\" (UniqueName: \"kubernetes.io/projected/99619e61-ede2-432a-a293-5fab27d7d60c-kube-api-access-tlbg7\") pod \"collect-profiles-29327550-zcp95\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.495169 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:00 crc kubenswrapper[4935]: I1005 08:30:00.958628 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95"] Oct 05 08:30:01 crc kubenswrapper[4935]: I1005 08:30:01.942816 4935 generic.go:334] "Generic (PLEG): container finished" podID="99619e61-ede2-432a-a293-5fab27d7d60c" containerID="0f7ff305a89f1eae0a5a300a0eeda823263feaf99b6ef15ed6d853f06474b825" exitCode=0 Oct 05 08:30:01 crc kubenswrapper[4935]: I1005 08:30:01.942925 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" event={"ID":"99619e61-ede2-432a-a293-5fab27d7d60c","Type":"ContainerDied","Data":"0f7ff305a89f1eae0a5a300a0eeda823263feaf99b6ef15ed6d853f06474b825"} Oct 05 08:30:01 crc kubenswrapper[4935]: I1005 08:30:01.943276 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" event={"ID":"99619e61-ede2-432a-a293-5fab27d7d60c","Type":"ContainerStarted","Data":"6c3c08522a1909b79e4918e8565dc405f05e446fe5231a90ced120e0b052dd59"} Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.275811 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.392932 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlbg7\" (UniqueName: \"kubernetes.io/projected/99619e61-ede2-432a-a293-5fab27d7d60c-kube-api-access-tlbg7\") pod \"99619e61-ede2-432a-a293-5fab27d7d60c\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.393007 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99619e61-ede2-432a-a293-5fab27d7d60c-config-volume\") pod \"99619e61-ede2-432a-a293-5fab27d7d60c\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.393053 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99619e61-ede2-432a-a293-5fab27d7d60c-secret-volume\") pod \"99619e61-ede2-432a-a293-5fab27d7d60c\" (UID: \"99619e61-ede2-432a-a293-5fab27d7d60c\") " Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.394071 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99619e61-ede2-432a-a293-5fab27d7d60c-config-volume" (OuterVolumeSpecName: "config-volume") pod "99619e61-ede2-432a-a293-5fab27d7d60c" (UID: "99619e61-ede2-432a-a293-5fab27d7d60c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.398500 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99619e61-ede2-432a-a293-5fab27d7d60c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "99619e61-ede2-432a-a293-5fab27d7d60c" (UID: "99619e61-ede2-432a-a293-5fab27d7d60c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.399408 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99619e61-ede2-432a-a293-5fab27d7d60c-kube-api-access-tlbg7" (OuterVolumeSpecName: "kube-api-access-tlbg7") pod "99619e61-ede2-432a-a293-5fab27d7d60c" (UID: "99619e61-ede2-432a-a293-5fab27d7d60c"). InnerVolumeSpecName "kube-api-access-tlbg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.495117 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlbg7\" (UniqueName: \"kubernetes.io/projected/99619e61-ede2-432a-a293-5fab27d7d60c-kube-api-access-tlbg7\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.495151 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99619e61-ede2-432a-a293-5fab27d7d60c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.495160 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99619e61-ede2-432a-a293-5fab27d7d60c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.963430 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" event={"ID":"99619e61-ede2-432a-a293-5fab27d7d60c","Type":"ContainerDied","Data":"6c3c08522a1909b79e4918e8565dc405f05e446fe5231a90ced120e0b052dd59"} Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.963495 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c3c08522a1909b79e4918e8565dc405f05e446fe5231a90ced120e0b052dd59" Oct 05 08:30:03 crc kubenswrapper[4935]: I1005 08:30:03.963519 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95" Oct 05 08:30:04 crc kubenswrapper[4935]: I1005 08:30:04.343807 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw"] Oct 05 08:30:04 crc kubenswrapper[4935]: I1005 08:30:04.349148 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-sxrpw"] Oct 05 08:30:04 crc kubenswrapper[4935]: I1005 08:30:04.796572 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d4f606e-3b56-46fa-b13c-0ca4a9c623e1" path="/var/lib/kubelet/pods/7d4f606e-3b56-46fa-b13c-0ca4a9c623e1/volumes" Oct 05 08:30:07 crc kubenswrapper[4935]: I1005 08:30:07.777024 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:30:07 crc kubenswrapper[4935]: E1005 08:30:07.777792 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:30:21 crc kubenswrapper[4935]: I1005 08:30:21.777424 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:30:21 crc kubenswrapper[4935]: E1005 08:30:21.778263 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:30:33 crc kubenswrapper[4935]: I1005 08:30:33.777650 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:30:33 crc kubenswrapper[4935]: E1005 08:30:33.778918 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:30:45 crc kubenswrapper[4935]: I1005 08:30:45.777724 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:30:45 crc kubenswrapper[4935]: E1005 08:30:45.778778 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.176429 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c55tj"] Oct 05 08:30:55 crc kubenswrapper[4935]: E1005 08:30:55.177435 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99619e61-ede2-432a-a293-5fab27d7d60c" containerName="collect-profiles" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.177452 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="99619e61-ede2-432a-a293-5fab27d7d60c" containerName="collect-profiles" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.177604 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="99619e61-ede2-432a-a293-5fab27d7d60c" containerName="collect-profiles" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.178320 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.180602 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.180734 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-sqbpb" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.180612 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.180692 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.183640 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.190538 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c55tj"] Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.334473 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-dns-svc\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.334585 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxjd2\" (UniqueName: \"kubernetes.io/projected/75406c9e-fb82-46e7-9f21-16c7cb657d84-kube-api-access-xxjd2\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.334668 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-config\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.435783 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxjd2\" (UniqueName: \"kubernetes.io/projected/75406c9e-fb82-46e7-9f21-16c7cb657d84-kube-api-access-xxjd2\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.436305 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-config\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.436817 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-lh64c"] Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.437237 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-config\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.450832 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-dns-svc\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.451953 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-dns-svc\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.456474 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.461290 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxjd2\" (UniqueName: \"kubernetes.io/projected/75406c9e-fb82-46e7-9f21-16c7cb657d84-kube-api-access-xxjd2\") pod \"dnsmasq-dns-78dbf54cbc-c55tj\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.479931 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-lh64c"] Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.498090 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.654048 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjcpr\" (UniqueName: \"kubernetes.io/projected/d029017e-6368-4136-b73a-1fac2ff17a1a-kube-api-access-cjcpr\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.654528 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-config\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.654561 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-dns-svc\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.756543 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjcpr\" (UniqueName: \"kubernetes.io/projected/d029017e-6368-4136-b73a-1fac2ff17a1a-kube-api-access-cjcpr\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.756608 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-config\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.756636 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-dns-svc\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.757548 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-dns-svc\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.757555 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-config\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.772802 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjcpr\" (UniqueName: \"kubernetes.io/projected/d029017e-6368-4136-b73a-1fac2ff17a1a-kube-api-access-cjcpr\") pod \"dnsmasq-dns-6cfd7b4c45-lh64c\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.876368 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:30:55 crc kubenswrapper[4935]: I1005 08:30:55.944426 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c55tj"] Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.289234 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-lh64c"] Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.306036 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.307530 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.310739 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.310777 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.310933 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.311028 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.311209 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7r8rg" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.337803 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.384031 4935 scope.go:117] "RemoveContainer" containerID="71fca8c49f608da923ef71e1b6d883361ba977867f3788d47af960feef5fb48a" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466595 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466641 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/121ca4c6-a507-46d8-949a-ffeb928d3ce5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466697 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466721 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466760 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466786 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466802 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466824 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/121ca4c6-a507-46d8-949a-ffeb928d3ce5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466844 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-487n2\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-kube-api-access-487n2\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.466844 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" event={"ID":"d029017e-6368-4136-b73a-1fac2ff17a1a","Type":"ContainerStarted","Data":"442f25cbc962e36bcaf5899d52b8a0ade6b3d35fc5256870db6beab1d0f060f2"} Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.468178 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" event={"ID":"75406c9e-fb82-46e7-9f21-16c7cb657d84","Type":"ContainerStarted","Data":"203be4f780e2f0fc57e83bc82c75bd75e398cfb5d437e0aff1701336773108e4"} Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.576715 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.576771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.576851 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.576976 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.577005 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.577036 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-487n2\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-kube-api-access-487n2\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.577060 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/121ca4c6-a507-46d8-949a-ffeb928d3ce5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.577172 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.577198 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/121ca4c6-a507-46d8-949a-ffeb928d3ce5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.578453 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.579360 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.579868 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.581657 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.581682 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/26bd9ab2ee5a1c8a155acf45110f5eead4872e3541b6a94d11ac934414657417/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.582182 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.583450 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.583618 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.584212 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/121ca4c6-a507-46d8-949a-ffeb928d3ce5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.584944 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.588451 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.588531 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mrrdw" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.588643 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.588808 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.591967 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.594705 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.599789 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/121ca4c6-a507-46d8-949a-ffeb928d3ce5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.617909 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-487n2\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-kube-api-access-487n2\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.673210 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"rabbitmq-server-0\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.780840 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.780924 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.780961 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.780996 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.781016 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.781096 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.781192 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.781228 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.781260 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxqt8\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-kube-api-access-qxqt8\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883364 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883418 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883471 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883514 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883566 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883591 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883642 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxqt8\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-kube-api-access-qxqt8\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883743 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.883966 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.884375 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.884562 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.885390 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.885722 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.887119 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.887569 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.887597 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c02f14fa6f28a7938e9ff24a5b29627787cdb2a2d00198967d957d270b3be65f/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.887661 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.895707 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.902735 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxqt8\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-kube-api-access-qxqt8\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.937690 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"rabbitmq-cell1-server-0\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.938418 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:30:56 crc kubenswrapper[4935]: I1005 08:30:56.993198 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.332391 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.489024 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"121ca4c6-a507-46d8-949a-ffeb928d3ce5","Type":"ContainerStarted","Data":"eb96acd20263166e2c7ae6b16747669eb25a8e134c9493732d14ba39a7c5b80a"} Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.648292 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.651578 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.653138 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.654186 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.654747 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.655293 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-dp9tc" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.655663 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.660343 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.662475 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.697689 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805273 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805339 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae622d60-4705-4af0-a65b-4c6654f825c7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805388 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805414 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-secrets\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805446 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4pgr\" (UniqueName: \"kubernetes.io/projected/ae622d60-4705-4af0-a65b-4c6654f825c7-kube-api-access-s4pgr\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805475 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805504 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805589 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.805628 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-20cf1ded-c067-4d0b-9011-a524acdd45b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20cf1ded-c067-4d0b-9011-a524acdd45b9\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907324 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907644 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae622d60-4705-4af0-a65b-4c6654f825c7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907686 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907703 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-secrets\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907727 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4pgr\" (UniqueName: \"kubernetes.io/projected/ae622d60-4705-4af0-a65b-4c6654f825c7-kube-api-access-s4pgr\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907745 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907769 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907824 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.907847 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-20cf1ded-c067-4d0b-9011-a524acdd45b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20cf1ded-c067-4d0b-9011-a524acdd45b9\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.908417 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae622d60-4705-4af0-a65b-4c6654f825c7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.909466 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.910616 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.920568 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae622d60-4705-4af0-a65b-4c6654f825c7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.926870 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-secrets\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.926902 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.929850 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.929879 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4pgr\" (UniqueName: \"kubernetes.io/projected/ae622d60-4705-4af0-a65b-4c6654f825c7-kube-api-access-s4pgr\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.929904 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-20cf1ded-c067-4d0b-9011-a524acdd45b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20cf1ded-c067-4d0b-9011-a524acdd45b9\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a5d41727c8954853d981f6364b9b6a9b16dc9e1e2efbdcf474f37c8a05ed757b/globalmount\"" pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.940284 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae622d60-4705-4af0-a65b-4c6654f825c7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.980366 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-20cf1ded-c067-4d0b-9011-a524acdd45b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-20cf1ded-c067-4d0b-9011-a524acdd45b9\") pod \"openstack-galera-0\" (UID: \"ae622d60-4705-4af0-a65b-4c6654f825c7\") " pod="openstack/openstack-galera-0" Oct 05 08:30:57 crc kubenswrapper[4935]: I1005 08:30:57.983229 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.098661 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.115989 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.119550 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.133214 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.133716 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-r9stv" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.214286 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aa569082-01a4-4f8c-97a1-665f630a6e53-kolla-config\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.214381 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aa569082-01a4-4f8c-97a1-665f630a6e53-config-data\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.214400 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb86r\" (UniqueName: \"kubernetes.io/projected/aa569082-01a4-4f8c-97a1-665f630a6e53-kube-api-access-xb86r\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.321721 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aa569082-01a4-4f8c-97a1-665f630a6e53-kolla-config\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.321801 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aa569082-01a4-4f8c-97a1-665f630a6e53-config-data\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.321819 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb86r\" (UniqueName: \"kubernetes.io/projected/aa569082-01a4-4f8c-97a1-665f630a6e53-kube-api-access-xb86r\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.322829 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aa569082-01a4-4f8c-97a1-665f630a6e53-kolla-config\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.323383 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aa569082-01a4-4f8c-97a1-665f630a6e53-config-data\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.338419 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb86r\" (UniqueName: \"kubernetes.io/projected/aa569082-01a4-4f8c-97a1-665f630a6e53-kube-api-access-xb86r\") pod \"memcached-0\" (UID: \"aa569082-01a4-4f8c-97a1-665f630a6e53\") " pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.472004 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 08:30:58 crc kubenswrapper[4935]: I1005 08:30:58.515448 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"213b1ad4-0921-4a45-aac5-4021f4a9f0b7","Type":"ContainerStarted","Data":"5daa404e6db8d07b03013417710b7eeca0ea91d84f1437f1a5a627b4eadc1cc3"} Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.164214 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.166712 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.168523 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.172162 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.172239 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.172693 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.172787 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5cs5h" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.323698 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.343842 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.343908 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.344246 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.344400 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.344545 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.344574 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-10845bf3-b80a-45ab-b0aa-8d64092b4170\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10845bf3-b80a-45ab-b0aa-8d64092b4170\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.344636 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfj57\" (UniqueName: \"kubernetes.io/projected/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-kube-api-access-cfj57\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.344962 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.345053 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.413736 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446629 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446712 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446758 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446784 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446837 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446868 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446907 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446928 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-10845bf3-b80a-45ab-b0aa-8d64092b4170\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10845bf3-b80a-45ab-b0aa-8d64092b4170\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.446949 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfj57\" (UniqueName: \"kubernetes.io/projected/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-kube-api-access-cfj57\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.447858 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.447876 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.449029 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.450592 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.450657 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-10845bf3-b80a-45ab-b0aa-8d64092b4170\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10845bf3-b80a-45ab-b0aa-8d64092b4170\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c477bf3472e2cf334bba5a3902053ffc2b8c8e3a8335aaea782729f3ca9c6d34/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.451518 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.453329 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.454287 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.462443 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfj57\" (UniqueName: \"kubernetes.io/projected/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-kube-api-access-cfj57\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.474810 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d558bd5-83a1-4c77-b06e-9c499e2e49b6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.484575 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-10845bf3-b80a-45ab-b0aa-8d64092b4170\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10845bf3-b80a-45ab-b0aa-8d64092b4170\") pod \"openstack-cell1-galera-0\" (UID: \"2d558bd5-83a1-4c77-b06e-9c499e2e49b6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.777541 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:30:59 crc kubenswrapper[4935]: E1005 08:30:59.777766 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:30:59 crc kubenswrapper[4935]: I1005 08:30:59.787039 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 08:31:08 crc kubenswrapper[4935]: I1005 08:31:08.606209 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"aa569082-01a4-4f8c-97a1-665f630a6e53","Type":"ContainerStarted","Data":"200846657e8ddb548095261ee292ed561955cfa3acb57c3c5977571d84011fa8"} Oct 05 08:31:08 crc kubenswrapper[4935]: I1005 08:31:08.613766 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae622d60-4705-4af0-a65b-4c6654f825c7","Type":"ContainerStarted","Data":"7cb6dd3d3907b201e9f07728cf54cecd33e10dcdd41f7b81473264fd39594891"} Oct 05 08:31:08 crc kubenswrapper[4935]: I1005 08:31:08.819301 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 08:31:09 crc kubenswrapper[4935]: I1005 08:31:09.623570 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d558bd5-83a1-4c77-b06e-9c499e2e49b6","Type":"ContainerStarted","Data":"a255dfbd752b961a9f0b225b9f6976d3b80fef9441ce4cda0bed3b3f37a0ac61"} Oct 05 08:31:09 crc kubenswrapper[4935]: I1005 08:31:09.625318 4935 generic.go:334] "Generic (PLEG): container finished" podID="d029017e-6368-4136-b73a-1fac2ff17a1a" containerID="afc3b8e9f1453d1c189ad92ef3f718d36342687109d0eeb6f4144a33339001de" exitCode=0 Oct 05 08:31:09 crc kubenswrapper[4935]: I1005 08:31:09.625509 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" event={"ID":"d029017e-6368-4136-b73a-1fac2ff17a1a","Type":"ContainerDied","Data":"afc3b8e9f1453d1c189ad92ef3f718d36342687109d0eeb6f4144a33339001de"} Oct 05 08:31:09 crc kubenswrapper[4935]: I1005 08:31:09.628032 4935 generic.go:334] "Generic (PLEG): container finished" podID="75406c9e-fb82-46e7-9f21-16c7cb657d84" containerID="4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141" exitCode=0 Oct 05 08:31:09 crc kubenswrapper[4935]: I1005 08:31:09.628087 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" event={"ID":"75406c9e-fb82-46e7-9f21-16c7cb657d84","Type":"ContainerDied","Data":"4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141"} Oct 05 08:31:10 crc kubenswrapper[4935]: I1005 08:31:10.636253 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"213b1ad4-0921-4a45-aac5-4021f4a9f0b7","Type":"ContainerStarted","Data":"169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4"} Oct 05 08:31:10 crc kubenswrapper[4935]: I1005 08:31:10.640956 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" event={"ID":"75406c9e-fb82-46e7-9f21-16c7cb657d84","Type":"ContainerStarted","Data":"c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e"} Oct 05 08:31:10 crc kubenswrapper[4935]: I1005 08:31:10.641088 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:31:10 crc kubenswrapper[4935]: I1005 08:31:10.643137 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"121ca4c6-a507-46d8-949a-ffeb928d3ce5","Type":"ContainerStarted","Data":"54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14"} Oct 05 08:31:10 crc kubenswrapper[4935]: I1005 08:31:10.648564 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" event={"ID":"d029017e-6368-4136-b73a-1fac2ff17a1a","Type":"ContainerStarted","Data":"5f6b7db7da25535b1f22d2c87e9d128be3ecc91306c460b0e7bfda5f8b1fb246"} Oct 05 08:31:10 crc kubenswrapper[4935]: I1005 08:31:10.648707 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:31:10 crc kubenswrapper[4935]: I1005 08:31:10.688726 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" podStartSLOduration=3.334670412 podStartE2EDuration="15.688708575s" podCreationTimestamp="2025-10-05 08:30:55 +0000 UTC" firstStartedPulling="2025-10-05 08:30:56.30746382 +0000 UTC m=+5890.190090280" lastFinishedPulling="2025-10-05 08:31:08.661501983 +0000 UTC m=+5902.544128443" observedRunningTime="2025-10-05 08:31:10.688255463 +0000 UTC m=+5904.570881923" watchObservedRunningTime="2025-10-05 08:31:10.688708575 +0000 UTC m=+5904.571335035" Oct 05 08:31:10 crc kubenswrapper[4935]: I1005 08:31:10.745286 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" podStartSLOduration=3.035559944 podStartE2EDuration="15.7452696s" podCreationTimestamp="2025-10-05 08:30:55 +0000 UTC" firstStartedPulling="2025-10-05 08:30:55.967004532 +0000 UTC m=+5889.849630992" lastFinishedPulling="2025-10-05 08:31:08.676714188 +0000 UTC m=+5902.559340648" observedRunningTime="2025-10-05 08:31:10.740450902 +0000 UTC m=+5904.623077362" watchObservedRunningTime="2025-10-05 08:31:10.7452696 +0000 UTC m=+5904.627896060" Oct 05 08:31:11 crc kubenswrapper[4935]: I1005 08:31:11.656189 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"aa569082-01a4-4f8c-97a1-665f630a6e53","Type":"ContainerStarted","Data":"2294768cfd089028b33450ed7cb52729cda66ae01434e36ac66d989c1464789e"} Oct 05 08:31:11 crc kubenswrapper[4935]: I1005 08:31:11.679104 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.500798539 podStartE2EDuration="13.679082523s" podCreationTimestamp="2025-10-05 08:30:58 +0000 UTC" firstStartedPulling="2025-10-05 08:31:07.783161667 +0000 UTC m=+5901.665788127" lastFinishedPulling="2025-10-05 08:31:10.961445651 +0000 UTC m=+5904.844072111" observedRunningTime="2025-10-05 08:31:11.673629908 +0000 UTC m=+5905.556256378" watchObservedRunningTime="2025-10-05 08:31:11.679082523 +0000 UTC m=+5905.561708983" Oct 05 08:31:12 crc kubenswrapper[4935]: I1005 08:31:12.662428 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 05 08:31:13 crc kubenswrapper[4935]: I1005 08:31:13.777172 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:31:13 crc kubenswrapper[4935]: E1005 08:31:13.777643 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:31:14 crc kubenswrapper[4935]: I1005 08:31:14.676330 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d558bd5-83a1-4c77-b06e-9c499e2e49b6","Type":"ContainerStarted","Data":"02d2c5371be4853c90fb763a988d322f818d63857b2bb75eb3289a2d9686026b"} Oct 05 08:31:14 crc kubenswrapper[4935]: I1005 08:31:14.678198 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae622d60-4705-4af0-a65b-4c6654f825c7","Type":"ContainerStarted","Data":"bccf147b4f1131b0b27d3bda0016d6ca6ae75d5215d2b57f17a0fa1a414dd0d8"} Oct 05 08:31:15 crc kubenswrapper[4935]: I1005 08:31:15.500152 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:31:15 crc kubenswrapper[4935]: I1005 08:31:15.878102 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:31:15 crc kubenswrapper[4935]: I1005 08:31:15.932073 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c55tj"] Oct 05 08:31:15 crc kubenswrapper[4935]: I1005 08:31:15.932271 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" podUID="75406c9e-fb82-46e7-9f21-16c7cb657d84" containerName="dnsmasq-dns" containerID="cri-o://c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e" gracePeriod=10 Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.421791 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.544186 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-config\") pod \"75406c9e-fb82-46e7-9f21-16c7cb657d84\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.544709 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-dns-svc\") pod \"75406c9e-fb82-46e7-9f21-16c7cb657d84\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.544751 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxjd2\" (UniqueName: \"kubernetes.io/projected/75406c9e-fb82-46e7-9f21-16c7cb657d84-kube-api-access-xxjd2\") pod \"75406c9e-fb82-46e7-9f21-16c7cb657d84\" (UID: \"75406c9e-fb82-46e7-9f21-16c7cb657d84\") " Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.553596 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75406c9e-fb82-46e7-9f21-16c7cb657d84-kube-api-access-xxjd2" (OuterVolumeSpecName: "kube-api-access-xxjd2") pod "75406c9e-fb82-46e7-9f21-16c7cb657d84" (UID: "75406c9e-fb82-46e7-9f21-16c7cb657d84"). InnerVolumeSpecName "kube-api-access-xxjd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.580666 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-config" (OuterVolumeSpecName: "config") pod "75406c9e-fb82-46e7-9f21-16c7cb657d84" (UID: "75406c9e-fb82-46e7-9f21-16c7cb657d84"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.584192 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "75406c9e-fb82-46e7-9f21-16c7cb657d84" (UID: "75406c9e-fb82-46e7-9f21-16c7cb657d84"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.647089 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.647118 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxjd2\" (UniqueName: \"kubernetes.io/projected/75406c9e-fb82-46e7-9f21-16c7cb657d84-kube-api-access-xxjd2\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.647130 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75406c9e-fb82-46e7-9f21-16c7cb657d84-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.693383 4935 generic.go:334] "Generic (PLEG): container finished" podID="75406c9e-fb82-46e7-9f21-16c7cb657d84" containerID="c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e" exitCode=0 Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.693436 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" event={"ID":"75406c9e-fb82-46e7-9f21-16c7cb657d84","Type":"ContainerDied","Data":"c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e"} Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.693458 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.693476 4935 scope.go:117] "RemoveContainer" containerID="c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.693465 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c55tj" event={"ID":"75406c9e-fb82-46e7-9f21-16c7cb657d84","Type":"ContainerDied","Data":"203be4f780e2f0fc57e83bc82c75bd75e398cfb5d437e0aff1701336773108e4"} Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.717058 4935 scope.go:117] "RemoveContainer" containerID="4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.726099 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c55tj"] Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.733963 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c55tj"] Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.746735 4935 scope.go:117] "RemoveContainer" containerID="c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e" Oct 05 08:31:16 crc kubenswrapper[4935]: E1005 08:31:16.747300 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e\": container with ID starting with c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e not found: ID does not exist" containerID="c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.747355 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e"} err="failed to get container status \"c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e\": rpc error: code = NotFound desc = could not find container \"c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e\": container with ID starting with c551e2e63d970cc0e7e3d5d099954d4d8c7c359cbb45553b5fdcd9f7f9e4590e not found: ID does not exist" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.747389 4935 scope.go:117] "RemoveContainer" containerID="4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141" Oct 05 08:31:16 crc kubenswrapper[4935]: E1005 08:31:16.747736 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141\": container with ID starting with 4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141 not found: ID does not exist" containerID="4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.747777 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141"} err="failed to get container status \"4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141\": rpc error: code = NotFound desc = could not find container \"4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141\": container with ID starting with 4ea417bc19cdf409d1adca3624cdf2995bc74214a61c8709d0e15a1971dc8141 not found: ID does not exist" Oct 05 08:31:16 crc kubenswrapper[4935]: I1005 08:31:16.786198 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75406c9e-fb82-46e7-9f21-16c7cb657d84" path="/var/lib/kubelet/pods/75406c9e-fb82-46e7-9f21-16c7cb657d84/volumes" Oct 05 08:31:17 crc kubenswrapper[4935]: I1005 08:31:17.706612 4935 generic.go:334] "Generic (PLEG): container finished" podID="2d558bd5-83a1-4c77-b06e-9c499e2e49b6" containerID="02d2c5371be4853c90fb763a988d322f818d63857b2bb75eb3289a2d9686026b" exitCode=0 Oct 05 08:31:17 crc kubenswrapper[4935]: I1005 08:31:17.706740 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d558bd5-83a1-4c77-b06e-9c499e2e49b6","Type":"ContainerDied","Data":"02d2c5371be4853c90fb763a988d322f818d63857b2bb75eb3289a2d9686026b"} Oct 05 08:31:17 crc kubenswrapper[4935]: I1005 08:31:17.713493 4935 generic.go:334] "Generic (PLEG): container finished" podID="ae622d60-4705-4af0-a65b-4c6654f825c7" containerID="bccf147b4f1131b0b27d3bda0016d6ca6ae75d5215d2b57f17a0fa1a414dd0d8" exitCode=0 Oct 05 08:31:17 crc kubenswrapper[4935]: I1005 08:31:17.713552 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae622d60-4705-4af0-a65b-4c6654f825c7","Type":"ContainerDied","Data":"bccf147b4f1131b0b27d3bda0016d6ca6ae75d5215d2b57f17a0fa1a414dd0d8"} Oct 05 08:31:18 crc kubenswrapper[4935]: I1005 08:31:18.473438 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 05 08:31:18 crc kubenswrapper[4935]: I1005 08:31:18.725314 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae622d60-4705-4af0-a65b-4c6654f825c7","Type":"ContainerStarted","Data":"48b60c1166ed8328bac43ff93b124c833b32bfb713363092ef0d086bb4a882de"} Oct 05 08:31:18 crc kubenswrapper[4935]: I1005 08:31:18.728263 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d558bd5-83a1-4c77-b06e-9c499e2e49b6","Type":"ContainerStarted","Data":"46e1671ced7b36fac8c01c17ea2b0b69b7d8fcbef37e4b737678fc6ba708373e"} Oct 05 08:31:18 crc kubenswrapper[4935]: I1005 08:31:18.747353 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.844254701 podStartE2EDuration="22.747318574s" podCreationTimestamp="2025-10-05 08:30:56 +0000 UTC" firstStartedPulling="2025-10-05 08:31:07.810519955 +0000 UTC m=+5901.693146415" lastFinishedPulling="2025-10-05 08:31:13.713583828 +0000 UTC m=+5907.596210288" observedRunningTime="2025-10-05 08:31:18.74263782 +0000 UTC m=+5912.625264280" watchObservedRunningTime="2025-10-05 08:31:18.747318574 +0000 UTC m=+5912.629945034" Oct 05 08:31:18 crc kubenswrapper[4935]: I1005 08:31:18.779132 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=15.903103029 podStartE2EDuration="20.77911459s" podCreationTimestamp="2025-10-05 08:30:58 +0000 UTC" firstStartedPulling="2025-10-05 08:31:08.831846805 +0000 UTC m=+5902.714473285" lastFinishedPulling="2025-10-05 08:31:13.707858386 +0000 UTC m=+5907.590484846" observedRunningTime="2025-10-05 08:31:18.77760196 +0000 UTC m=+5912.660228430" watchObservedRunningTime="2025-10-05 08:31:18.77911459 +0000 UTC m=+5912.661741060" Oct 05 08:31:19 crc kubenswrapper[4935]: I1005 08:31:19.788076 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 05 08:31:19 crc kubenswrapper[4935]: I1005 08:31:19.788131 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 05 08:31:23 crc kubenswrapper[4935]: I1005 08:31:23.845503 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 05 08:31:23 crc kubenswrapper[4935]: I1005 08:31:23.915120 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 05 08:31:25 crc kubenswrapper[4935]: I1005 08:31:25.777490 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:31:26 crc kubenswrapper[4935]: I1005 08:31:26.791562 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"c98eb32c261a6373bd85357643e260ea4fb13065049a0fdbd29241582949a384"} Oct 05 08:31:27 crc kubenswrapper[4935]: I1005 08:31:27.984275 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 05 08:31:27 crc kubenswrapper[4935]: I1005 08:31:27.984618 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 05 08:31:28 crc kubenswrapper[4935]: I1005 08:31:28.047608 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 05 08:31:28 crc kubenswrapper[4935]: I1005 08:31:28.866408 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 05 08:31:41 crc kubenswrapper[4935]: I1005 08:31:41.927673 4935 generic.go:334] "Generic (PLEG): container finished" podID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerID="169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4" exitCode=0 Oct 05 08:31:41 crc kubenswrapper[4935]: I1005 08:31:41.927751 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"213b1ad4-0921-4a45-aac5-4021f4a9f0b7","Type":"ContainerDied","Data":"169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4"} Oct 05 08:31:41 crc kubenswrapper[4935]: I1005 08:31:41.931469 4935 generic.go:334] "Generic (PLEG): container finished" podID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerID="54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14" exitCode=0 Oct 05 08:31:41 crc kubenswrapper[4935]: I1005 08:31:41.931513 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"121ca4c6-a507-46d8-949a-ffeb928d3ce5","Type":"ContainerDied","Data":"54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14"} Oct 05 08:31:42 crc kubenswrapper[4935]: I1005 08:31:42.943757 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"121ca4c6-a507-46d8-949a-ffeb928d3ce5","Type":"ContainerStarted","Data":"07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad"} Oct 05 08:31:42 crc kubenswrapper[4935]: I1005 08:31:42.944600 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 05 08:31:42 crc kubenswrapper[4935]: I1005 08:31:42.947256 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"213b1ad4-0921-4a45-aac5-4021f4a9f0b7","Type":"ContainerStarted","Data":"1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c"} Oct 05 08:31:42 crc kubenswrapper[4935]: I1005 08:31:42.947880 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:31:42 crc kubenswrapper[4935]: I1005 08:31:42.975176 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.715492097 podStartE2EDuration="47.975149267s" podCreationTimestamp="2025-10-05 08:30:55 +0000 UTC" firstStartedPulling="2025-10-05 08:30:57.353263592 +0000 UTC m=+5891.235890052" lastFinishedPulling="2025-10-05 08:31:08.612920772 +0000 UTC m=+5902.495547222" observedRunningTime="2025-10-05 08:31:42.96250225 +0000 UTC m=+5936.845128770" watchObservedRunningTime="2025-10-05 08:31:42.975149267 +0000 UTC m=+5936.857775767" Oct 05 08:31:56 crc kubenswrapper[4935]: I1005 08:31:56.943289 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 05 08:31:56 crc kubenswrapper[4935]: I1005 08:31:56.979863 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.081124237 podStartE2EDuration="1m1.979837574s" podCreationTimestamp="2025-10-05 08:30:55 +0000 UTC" firstStartedPulling="2025-10-05 08:30:57.717582744 +0000 UTC m=+5891.600209204" lastFinishedPulling="2025-10-05 08:31:08.616296081 +0000 UTC m=+5902.498922541" observedRunningTime="2025-10-05 08:31:43.003727387 +0000 UTC m=+5936.886353847" watchObservedRunningTime="2025-10-05 08:31:56.979837574 +0000 UTC m=+5950.862464034" Oct 05 08:31:56 crc kubenswrapper[4935]: I1005 08:31:56.997360 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.475123 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rlsd8"] Oct 05 08:32:03 crc kubenswrapper[4935]: E1005 08:32:03.475988 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75406c9e-fb82-46e7-9f21-16c7cb657d84" containerName="dnsmasq-dns" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.476006 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="75406c9e-fb82-46e7-9f21-16c7cb657d84" containerName="dnsmasq-dns" Oct 05 08:32:03 crc kubenswrapper[4935]: E1005 08:32:03.476018 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75406c9e-fb82-46e7-9f21-16c7cb657d84" containerName="init" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.476024 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="75406c9e-fb82-46e7-9f21-16c7cb657d84" containerName="init" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.476196 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="75406c9e-fb82-46e7-9f21-16c7cb657d84" containerName="dnsmasq-dns" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.477282 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.541999 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rlsd8"] Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.581180 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-n6qwp"] Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.583724 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.594483 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-n6qwp"] Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.637753 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-catalog-content\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.637844 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cngss\" (UniqueName: \"kubernetes.io/projected/39ad8b37-a17c-4a01-8c88-7072f09c42eb-kube-api-access-cngss\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.637973 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-utilities\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.739870 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-catalog-content\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.740031 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j585\" (UniqueName: \"kubernetes.io/projected/60e2b0c1-83ce-4061-9baa-351e27b30b09-kube-api-access-8j585\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.740070 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cngss\" (UniqueName: \"kubernetes.io/projected/39ad8b37-a17c-4a01-8c88-7072f09c42eb-kube-api-access-cngss\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.740091 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-utilities\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.740117 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-dns-svc\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.740134 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-config\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.740436 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-catalog-content\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.740589 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-utilities\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.761640 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cngss\" (UniqueName: \"kubernetes.io/projected/39ad8b37-a17c-4a01-8c88-7072f09c42eb-kube-api-access-cngss\") pod \"redhat-operators-rlsd8\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.793170 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.841422 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j585\" (UniqueName: \"kubernetes.io/projected/60e2b0c1-83ce-4061-9baa-351e27b30b09-kube-api-access-8j585\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.841812 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-dns-svc\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.841839 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-config\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.842698 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-dns-svc\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.842770 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-config\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.857469 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j585\" (UniqueName: \"kubernetes.io/projected/60e2b0c1-83ce-4061-9baa-351e27b30b09-kube-api-access-8j585\") pod \"dnsmasq-dns-96d5866c7-n6qwp\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:03 crc kubenswrapper[4935]: I1005 08:32:03.898446 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:04 crc kubenswrapper[4935]: I1005 08:32:04.293466 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:32:04 crc kubenswrapper[4935]: I1005 08:32:04.342840 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rlsd8"] Oct 05 08:32:04 crc kubenswrapper[4935]: W1005 08:32:04.348881 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39ad8b37_a17c_4a01_8c88_7072f09c42eb.slice/crio-ab57cc8cc93663c719fde1bfaa314b5b8ca5de283c02f8a1c52ca8f0e7e23e38 WatchSource:0}: Error finding container ab57cc8cc93663c719fde1bfaa314b5b8ca5de283c02f8a1c52ca8f0e7e23e38: Status 404 returned error can't find the container with id ab57cc8cc93663c719fde1bfaa314b5b8ca5de283c02f8a1c52ca8f0e7e23e38 Oct 05 08:32:04 crc kubenswrapper[4935]: I1005 08:32:04.443240 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-n6qwp"] Oct 05 08:32:04 crc kubenswrapper[4935]: W1005 08:32:04.448754 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60e2b0c1_83ce_4061_9baa_351e27b30b09.slice/crio-7a8b8eac8db63dd351b0513b7831be6361336d3fd2202820d424bb74c726de7e WatchSource:0}: Error finding container 7a8b8eac8db63dd351b0513b7831be6361336d3fd2202820d424bb74c726de7e: Status 404 returned error can't find the container with id 7a8b8eac8db63dd351b0513b7831be6361336d3fd2202820d424bb74c726de7e Oct 05 08:32:04 crc kubenswrapper[4935]: I1005 08:32:04.973692 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:32:05 crc kubenswrapper[4935]: I1005 08:32:05.135574 4935 generic.go:334] "Generic (PLEG): container finished" podID="60e2b0c1-83ce-4061-9baa-351e27b30b09" containerID="97d71a2bc65cbd504001d9e7d84a34319439faf6f51a9aacf005b96ab7e98856" exitCode=0 Oct 05 08:32:05 crc kubenswrapper[4935]: I1005 08:32:05.135662 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" event={"ID":"60e2b0c1-83ce-4061-9baa-351e27b30b09","Type":"ContainerDied","Data":"97d71a2bc65cbd504001d9e7d84a34319439faf6f51a9aacf005b96ab7e98856"} Oct 05 08:32:05 crc kubenswrapper[4935]: I1005 08:32:05.135690 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" event={"ID":"60e2b0c1-83ce-4061-9baa-351e27b30b09","Type":"ContainerStarted","Data":"7a8b8eac8db63dd351b0513b7831be6361336d3fd2202820d424bb74c726de7e"} Oct 05 08:32:05 crc kubenswrapper[4935]: I1005 08:32:05.137386 4935 generic.go:334] "Generic (PLEG): container finished" podID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerID="ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44" exitCode=0 Oct 05 08:32:05 crc kubenswrapper[4935]: I1005 08:32:05.137433 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlsd8" event={"ID":"39ad8b37-a17c-4a01-8c88-7072f09c42eb","Type":"ContainerDied","Data":"ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44"} Oct 05 08:32:05 crc kubenswrapper[4935]: I1005 08:32:05.137460 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlsd8" event={"ID":"39ad8b37-a17c-4a01-8c88-7072f09c42eb","Type":"ContainerStarted","Data":"ab57cc8cc93663c719fde1bfaa314b5b8ca5de283c02f8a1c52ca8f0e7e23e38"} Oct 05 08:32:05 crc kubenswrapper[4935]: I1005 08:32:05.139211 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:32:06 crc kubenswrapper[4935]: I1005 08:32:06.121465 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerName="rabbitmq" containerID="cri-o://07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad" gracePeriod=604799 Oct 05 08:32:06 crc kubenswrapper[4935]: I1005 08:32:06.147244 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" event={"ID":"60e2b0c1-83ce-4061-9baa-351e27b30b09","Type":"ContainerStarted","Data":"2b4f1aff70ad057031684c6d6c77872d598fe45472224d32a3ef39d7a4bd3297"} Oct 05 08:32:06 crc kubenswrapper[4935]: I1005 08:32:06.147403 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:06 crc kubenswrapper[4935]: I1005 08:32:06.149391 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlsd8" event={"ID":"39ad8b37-a17c-4a01-8c88-7072f09c42eb","Type":"ContainerStarted","Data":"d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5"} Oct 05 08:32:06 crc kubenswrapper[4935]: I1005 08:32:06.167398 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" podStartSLOduration=3.167381017 podStartE2EDuration="3.167381017s" podCreationTimestamp="2025-10-05 08:32:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:32:06.165014284 +0000 UTC m=+5960.047640744" watchObservedRunningTime="2025-10-05 08:32:06.167381017 +0000 UTC m=+5960.050007477" Oct 05 08:32:06 crc kubenswrapper[4935]: I1005 08:32:06.737218 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerName="rabbitmq" containerID="cri-o://1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c" gracePeriod=604799 Oct 05 08:32:06 crc kubenswrapper[4935]: I1005 08:32:06.940440 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.247:5672: connect: connection refused" Oct 05 08:32:06 crc kubenswrapper[4935]: I1005 08:32:06.994632 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.248:5672: connect: connection refused" Oct 05 08:32:07 crc kubenswrapper[4935]: I1005 08:32:07.159773 4935 generic.go:334] "Generic (PLEG): container finished" podID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerID="d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5" exitCode=0 Oct 05 08:32:07 crc kubenswrapper[4935]: I1005 08:32:07.159930 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlsd8" event={"ID":"39ad8b37-a17c-4a01-8c88-7072f09c42eb","Type":"ContainerDied","Data":"d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5"} Oct 05 08:32:08 crc kubenswrapper[4935]: I1005 08:32:08.170710 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlsd8" event={"ID":"39ad8b37-a17c-4a01-8c88-7072f09c42eb","Type":"ContainerStarted","Data":"700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42"} Oct 05 08:32:08 crc kubenswrapper[4935]: I1005 08:32:08.207072 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rlsd8" podStartSLOduration=2.722030669 podStartE2EDuration="5.20705115s" podCreationTimestamp="2025-10-05 08:32:03 +0000 UTC" firstStartedPulling="2025-10-05 08:32:05.139021439 +0000 UTC m=+5959.021647899" lastFinishedPulling="2025-10-05 08:32:07.62404188 +0000 UTC m=+5961.506668380" observedRunningTime="2025-10-05 08:32:08.198295427 +0000 UTC m=+5962.080921917" watchObservedRunningTime="2025-10-05 08:32:08.20705115 +0000 UTC m=+5962.089677610" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.710949 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.913883 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-erlang-cookie\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.913961 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/121ca4c6-a507-46d8-949a-ffeb928d3ce5-erlang-cookie-secret\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.913992 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-plugins\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.914030 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-plugins-conf\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.914056 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/121ca4c6-a507-46d8-949a-ffeb928d3ce5-pod-info\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.914268 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-server-conf\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.914463 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.914586 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-487n2\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-kube-api-access-487n2\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.914682 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-confd\") pod \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\" (UID: \"121ca4c6-a507-46d8-949a-ffeb928d3ce5\") " Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.914839 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.915252 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.915243 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.915505 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.920480 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/121ca4c6-a507-46d8-949a-ffeb928d3ce5-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.921292 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/121ca4c6-a507-46d8-949a-ffeb928d3ce5-pod-info" (OuterVolumeSpecName: "pod-info") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.931549 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-kube-api-access-487n2" (OuterVolumeSpecName: "kube-api-access-487n2") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "kube-api-access-487n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.934816 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088" (OuterVolumeSpecName: "persistence") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.941335 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-server-conf" (OuterVolumeSpecName: "server-conf") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:32:12 crc kubenswrapper[4935]: I1005 08:32:12.975149 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "121ca4c6-a507-46d8-949a-ffeb928d3ce5" (UID: "121ca4c6-a507-46d8-949a-ffeb928d3ce5"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.016239 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-487n2\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-kube-api-access-487n2\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.016274 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.016284 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/121ca4c6-a507-46d8-949a-ffeb928d3ce5-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.016293 4935 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/121ca4c6-a507-46d8-949a-ffeb928d3ce5-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.016302 4935 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.016313 4935 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/121ca4c6-a507-46d8-949a-ffeb928d3ce5-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.016322 4935 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/121ca4c6-a507-46d8-949a-ffeb928d3ce5-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.016360 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") on node \"crc\" " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.032982 4935 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.033158 4935 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088") on node "crc" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.117806 4935 reconciler_common.go:293] "Volume detached for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.215063 4935 generic.go:334] "Generic (PLEG): container finished" podID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerID="07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad" exitCode=0 Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.215108 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"121ca4c6-a507-46d8-949a-ffeb928d3ce5","Type":"ContainerDied","Data":"07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad"} Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.215135 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"121ca4c6-a507-46d8-949a-ffeb928d3ce5","Type":"ContainerDied","Data":"eb96acd20263166e2c7ae6b16747669eb25a8e134c9493732d14ba39a7c5b80a"} Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.215151 4935 scope.go:117] "RemoveContainer" containerID="07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.215240 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.258563 4935 scope.go:117] "RemoveContainer" containerID="54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.276916 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.285718 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.300183 4935 scope.go:117] "RemoveContainer" containerID="07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad" Oct 05 08:32:13 crc kubenswrapper[4935]: E1005 08:32:13.300930 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad\": container with ID starting with 07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad not found: ID does not exist" containerID="07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.301001 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad"} err="failed to get container status \"07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad\": rpc error: code = NotFound desc = could not find container \"07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad\": container with ID starting with 07d3fd3f26f70ad0fdcb8fd95b91a32813980d84e295a7c8b7fac7ab4a9366ad not found: ID does not exist" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.301042 4935 scope.go:117] "RemoveContainer" containerID="54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14" Oct 05 08:32:13 crc kubenswrapper[4935]: E1005 08:32:13.301505 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14\": container with ID starting with 54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14 not found: ID does not exist" containerID="54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.301652 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14"} err="failed to get container status \"54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14\": rpc error: code = NotFound desc = could not find container \"54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14\": container with ID starting with 54c111737f10461a3498e2176e7247ab44e535c080b7bc6b7a818198a39b7a14 not found: ID does not exist" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.346395 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:32:13 crc kubenswrapper[4935]: E1005 08:32:13.347066 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerName="setup-container" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.347769 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerName="setup-container" Oct 05 08:32:13 crc kubenswrapper[4935]: E1005 08:32:13.347871 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerName="rabbitmq" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.347965 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerName="rabbitmq" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.348279 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" containerName="rabbitmq" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.349565 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.353155 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.353385 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.353627 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.353787 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.359496 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7r8rg" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.365567 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.534108 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.534159 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.534179 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.534766 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.534944 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.535025 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.535145 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j96k\" (UniqueName: \"kubernetes.io/projected/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-kube-api-access-9j96k\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.535208 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.535289 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636632 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636698 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636722 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636740 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636810 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636835 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636869 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j96k\" (UniqueName: \"kubernetes.io/projected/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-kube-api-access-9j96k\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.636886 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.637751 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.638010 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.638959 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.639644 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.640313 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.640340 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/26bd9ab2ee5a1c8a155acf45110f5eead4872e3541b6a94d11ac934414657417/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.643570 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.643824 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.658518 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.659472 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j96k\" (UniqueName: \"kubernetes.io/projected/f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377-kube-api-access-9j96k\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.672566 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5c85f80e-e376-4b9a-8fcc-6ed532758088\") pod \"rabbitmq-server-0\" (UID: \"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377\") " pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.707612 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.728881 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739469 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-plugins-conf\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739540 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-plugins\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739579 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-erlang-cookie\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739619 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-erlang-cookie-secret\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739673 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-server-conf\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739744 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-confd\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739782 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxqt8\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-kube-api-access-qxqt8\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739938 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.739997 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.740094 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-pod-info\") pod \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\" (UID: \"213b1ad4-0921-4a45-aac5-4021f4a9f0b7\") " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.740346 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.740453 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.744018 4935 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.744060 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.744082 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.744068 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.744623 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-kube-api-access-qxqt8" (OuterVolumeSpecName: "kube-api-access-qxqt8") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "kube-api-access-qxqt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.745942 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-pod-info" (OuterVolumeSpecName: "pod-info") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.762979 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c" (OuterVolumeSpecName: "persistence") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.788300 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-server-conf" (OuterVolumeSpecName: "server-conf") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.793539 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.793591 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.842843 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "213b1ad4-0921-4a45-aac5-4021f4a9f0b7" (UID: "213b1ad4-0921-4a45-aac5-4021f4a9f0b7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.847137 4935 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.847169 4935 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.847184 4935 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.847196 4935 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.847211 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxqt8\" (UniqueName: \"kubernetes.io/projected/213b1ad4-0921-4a45-aac5-4021f4a9f0b7-kube-api-access-qxqt8\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.847240 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") on node \"crc\" " Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.855605 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.872284 4935 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.872419 4935 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c") on node "crc" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.900717 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.949099 4935 reconciler_common.go:293] "Volume detached for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.956422 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-lh64c"] Oct 05 08:32:13 crc kubenswrapper[4935]: I1005 08:32:13.956856 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" podUID="d029017e-6368-4136-b73a-1fac2ff17a1a" containerName="dnsmasq-dns" containerID="cri-o://5f6b7db7da25535b1f22d2c87e9d128be3ecc91306c460b0e7bfda5f8b1fb246" gracePeriod=10 Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.205367 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:32:14 crc kubenswrapper[4935]: W1005 08:32:14.211881 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7d76a16_e4d9_4e84_81a8_fc4bfdfc0377.slice/crio-654e51e46a990ee0f92be8113b1aef88eba499b30426807e1e4e6bc3cf146076 WatchSource:0}: Error finding container 654e51e46a990ee0f92be8113b1aef88eba499b30426807e1e4e6bc3cf146076: Status 404 returned error can't find the container with id 654e51e46a990ee0f92be8113b1aef88eba499b30426807e1e4e6bc3cf146076 Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.233717 4935 generic.go:334] "Generic (PLEG): container finished" podID="d029017e-6368-4136-b73a-1fac2ff17a1a" containerID="5f6b7db7da25535b1f22d2c87e9d128be3ecc91306c460b0e7bfda5f8b1fb246" exitCode=0 Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.233823 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" event={"ID":"d029017e-6368-4136-b73a-1fac2ff17a1a","Type":"ContainerDied","Data":"5f6b7db7da25535b1f22d2c87e9d128be3ecc91306c460b0e7bfda5f8b1fb246"} Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.238952 4935 generic.go:334] "Generic (PLEG): container finished" podID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerID="1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c" exitCode=0 Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.239118 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.239701 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"213b1ad4-0921-4a45-aac5-4021f4a9f0b7","Type":"ContainerDied","Data":"1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c"} Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.239764 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"213b1ad4-0921-4a45-aac5-4021f4a9f0b7","Type":"ContainerDied","Data":"5daa404e6db8d07b03013417710b7eeca0ea91d84f1437f1a5a627b4eadc1cc3"} Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.239787 4935 scope.go:117] "RemoveContainer" containerID="1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.252052 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377","Type":"ContainerStarted","Data":"654e51e46a990ee0f92be8113b1aef88eba499b30426807e1e4e6bc3cf146076"} Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.299043 4935 scope.go:117] "RemoveContainer" containerID="169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.305189 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.311339 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.319160 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.334737 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:32:14 crc kubenswrapper[4935]: E1005 08:32:14.335274 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerName="setup-container" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.335292 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerName="setup-container" Oct 05 08:32:14 crc kubenswrapper[4935]: E1005 08:32:14.335344 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerName="rabbitmq" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.335351 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerName="rabbitmq" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.335571 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" containerName="rabbitmq" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.336678 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.339005 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.339255 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.339424 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.339532 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mrrdw" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.343920 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.346828 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.356611 4935 scope.go:117] "RemoveContainer" containerID="1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c" Oct 05 08:32:14 crc kubenswrapper[4935]: E1005 08:32:14.357121 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c\": container with ID starting with 1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c not found: ID does not exist" containerID="1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.357147 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c"} err="failed to get container status \"1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c\": rpc error: code = NotFound desc = could not find container \"1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c\": container with ID starting with 1a5ffbac2e869f0959b74888c28be3ae01efd8af1ad9399c89a9960e79b10e0c not found: ID does not exist" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.357166 4935 scope.go:117] "RemoveContainer" containerID="169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4" Oct 05 08:32:14 crc kubenswrapper[4935]: E1005 08:32:14.357424 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4\": container with ID starting with 169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4 not found: ID does not exist" containerID="169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.357466 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4"} err="failed to get container status \"169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4\": rpc error: code = NotFound desc = could not find container \"169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4\": container with ID starting with 169da75fde483397cbf39d09ca28fde74797e329adcb902c00c22f2a5aa1c1e4 not found: ID does not exist" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.378353 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rlsd8"] Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.400589 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.458191 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.458849 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.458989 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.459069 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.459192 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2vvb\" (UniqueName: \"kubernetes.io/projected/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-kube-api-access-t2vvb\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.459276 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.459373 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.459478 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.459549 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.560561 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjcpr\" (UniqueName: \"kubernetes.io/projected/d029017e-6368-4136-b73a-1fac2ff17a1a-kube-api-access-cjcpr\") pod \"d029017e-6368-4136-b73a-1fac2ff17a1a\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.560723 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-dns-svc\") pod \"d029017e-6368-4136-b73a-1fac2ff17a1a\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.560757 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-config\") pod \"d029017e-6368-4136-b73a-1fac2ff17a1a\" (UID: \"d029017e-6368-4136-b73a-1fac2ff17a1a\") " Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.560992 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561022 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561043 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561063 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561091 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561113 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561573 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561740 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2vvb\" (UniqueName: \"kubernetes.io/projected/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-kube-api-access-t2vvb\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561782 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.561818 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.562033 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.562351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.562546 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.564378 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.564408 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c02f14fa6f28a7938e9ff24a5b29627787cdb2a2d00198967d957d270b3be65f/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.567086 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d029017e-6368-4136-b73a-1fac2ff17a1a-kube-api-access-cjcpr" (OuterVolumeSpecName: "kube-api-access-cjcpr") pod "d029017e-6368-4136-b73a-1fac2ff17a1a" (UID: "d029017e-6368-4136-b73a-1fac2ff17a1a"). InnerVolumeSpecName "kube-api-access-cjcpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.567594 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.571011 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.575126 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.589060 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2vvb\" (UniqueName: \"kubernetes.io/projected/0a21d4cb-88f1-4bec-ad29-97e503d2dad5-kube-api-access-t2vvb\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.606369 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-config" (OuterVolumeSpecName: "config") pod "d029017e-6368-4136-b73a-1fac2ff17a1a" (UID: "d029017e-6368-4136-b73a-1fac2ff17a1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.613223 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-736bb8c5-af67-4e02-8cd1-339f9b97065c\") pod \"rabbitmq-cell1-server-0\" (UID: \"0a21d4cb-88f1-4bec-ad29-97e503d2dad5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.614081 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d029017e-6368-4136-b73a-1fac2ff17a1a" (UID: "d029017e-6368-4136-b73a-1fac2ff17a1a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.663299 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjcpr\" (UniqueName: \"kubernetes.io/projected/d029017e-6368-4136-b73a-1fac2ff17a1a-kube-api-access-cjcpr\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.663468 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.663552 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d029017e-6368-4136-b73a-1fac2ff17a1a-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.712079 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.789926 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="121ca4c6-a507-46d8-949a-ffeb928d3ce5" path="/var/lib/kubelet/pods/121ca4c6-a507-46d8-949a-ffeb928d3ce5/volumes" Oct 05 08:32:14 crc kubenswrapper[4935]: I1005 08:32:14.791456 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="213b1ad4-0921-4a45-aac5-4021f4a9f0b7" path="/var/lib/kubelet/pods/213b1ad4-0921-4a45-aac5-4021f4a9f0b7/volumes" Oct 05 08:32:14 crc kubenswrapper[4935]: E1005 08:32:14.955880 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd029017e_6368_4136_b73a_1fac2ff17a1a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd029017e_6368_4136_b73a_1fac2ff17a1a.slice/crio-442f25cbc962e36bcaf5899d52b8a0ade6b3d35fc5256870db6beab1d0f060f2\": RecentStats: unable to find data in memory cache]" Oct 05 08:32:15 crc kubenswrapper[4935]: I1005 08:32:15.268597 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" Oct 05 08:32:15 crc kubenswrapper[4935]: I1005 08:32:15.270100 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-lh64c" event={"ID":"d029017e-6368-4136-b73a-1fac2ff17a1a","Type":"ContainerDied","Data":"442f25cbc962e36bcaf5899d52b8a0ade6b3d35fc5256870db6beab1d0f060f2"} Oct 05 08:32:15 crc kubenswrapper[4935]: I1005 08:32:15.270166 4935 scope.go:117] "RemoveContainer" containerID="5f6b7db7da25535b1f22d2c87e9d128be3ecc91306c460b0e7bfda5f8b1fb246" Oct 05 08:32:15 crc kubenswrapper[4935]: I1005 08:32:15.270255 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:32:15 crc kubenswrapper[4935]: I1005 08:32:15.292509 4935 scope.go:117] "RemoveContainer" containerID="afc3b8e9f1453d1c189ad92ef3f718d36342687109d0eeb6f4144a33339001de" Oct 05 08:32:15 crc kubenswrapper[4935]: I1005 08:32:15.299790 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-lh64c"] Oct 05 08:32:15 crc kubenswrapper[4935]: I1005 08:32:15.308086 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-lh64c"] Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.285288 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377","Type":"ContainerStarted","Data":"d0ddee54ea9f08be12ab9f6ce977b9c9e03523d87464e394ab6efe2875687f79"} Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.287674 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0a21d4cb-88f1-4bec-ad29-97e503d2dad5","Type":"ContainerStarted","Data":"05be7655462d9d2000edaf5f38c57e7f92ea339f59217048b74dd3b154808ec3"} Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.290082 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rlsd8" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerName="registry-server" containerID="cri-o://700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42" gracePeriod=2 Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.724722 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.786371 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d029017e-6368-4136-b73a-1fac2ff17a1a" path="/var/lib/kubelet/pods/d029017e-6368-4136-b73a-1fac2ff17a1a/volumes" Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.904306 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-utilities\") pod \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.904390 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-catalog-content\") pod \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.904505 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cngss\" (UniqueName: \"kubernetes.io/projected/39ad8b37-a17c-4a01-8c88-7072f09c42eb-kube-api-access-cngss\") pod \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\" (UID: \"39ad8b37-a17c-4a01-8c88-7072f09c42eb\") " Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.905411 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-utilities" (OuterVolumeSpecName: "utilities") pod "39ad8b37-a17c-4a01-8c88-7072f09c42eb" (UID: "39ad8b37-a17c-4a01-8c88-7072f09c42eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:32:16 crc kubenswrapper[4935]: I1005 08:32:16.912141 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ad8b37-a17c-4a01-8c88-7072f09c42eb-kube-api-access-cngss" (OuterVolumeSpecName: "kube-api-access-cngss") pod "39ad8b37-a17c-4a01-8c88-7072f09c42eb" (UID: "39ad8b37-a17c-4a01-8c88-7072f09c42eb"). InnerVolumeSpecName "kube-api-access-cngss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.007037 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.007070 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cngss\" (UniqueName: \"kubernetes.io/projected/39ad8b37-a17c-4a01-8c88-7072f09c42eb-kube-api-access-cngss\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.301563 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0a21d4cb-88f1-4bec-ad29-97e503d2dad5","Type":"ContainerStarted","Data":"da588cb280e0a39caeb1ec3b876aa1e82fea050b29a1e5b103216f6951acf70d"} Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.304536 4935 generic.go:334] "Generic (PLEG): container finished" podID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerID="700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42" exitCode=0 Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.305175 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlsd8" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.305343 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlsd8" event={"ID":"39ad8b37-a17c-4a01-8c88-7072f09c42eb","Type":"ContainerDied","Data":"700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42"} Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.305377 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlsd8" event={"ID":"39ad8b37-a17c-4a01-8c88-7072f09c42eb","Type":"ContainerDied","Data":"ab57cc8cc93663c719fde1bfaa314b5b8ca5de283c02f8a1c52ca8f0e7e23e38"} Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.305398 4935 scope.go:117] "RemoveContainer" containerID="700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.344899 4935 scope.go:117] "RemoveContainer" containerID="d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.371714 4935 scope.go:117] "RemoveContainer" containerID="ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.390387 4935 scope.go:117] "RemoveContainer" containerID="700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42" Oct 05 08:32:17 crc kubenswrapper[4935]: E1005 08:32:17.391029 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42\": container with ID starting with 700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42 not found: ID does not exist" containerID="700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.391057 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42"} err="failed to get container status \"700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42\": rpc error: code = NotFound desc = could not find container \"700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42\": container with ID starting with 700a6909494977d4e6e0ad6cd6b5a6bc2a05020cf36cd3af5a87dc4859dcee42 not found: ID does not exist" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.391075 4935 scope.go:117] "RemoveContainer" containerID="d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5" Oct 05 08:32:17 crc kubenswrapper[4935]: E1005 08:32:17.391478 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5\": container with ID starting with d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5 not found: ID does not exist" containerID="d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.391505 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5"} err="failed to get container status \"d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5\": rpc error: code = NotFound desc = could not find container \"d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5\": container with ID starting with d6b2f2e3ca20307c88ddc68b426b9242a8248c2e423e7c77c6da5a813e3247b5 not found: ID does not exist" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.391523 4935 scope.go:117] "RemoveContainer" containerID="ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44" Oct 05 08:32:17 crc kubenswrapper[4935]: E1005 08:32:17.391815 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44\": container with ID starting with ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44 not found: ID does not exist" containerID="ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.391839 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44"} err="failed to get container status \"ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44\": rpc error: code = NotFound desc = could not find container \"ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44\": container with ID starting with ecf9bbf7182d90429754be193ab1b6b9c7090f225cedff747ddcf57997845d44 not found: ID does not exist" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.897934 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39ad8b37-a17c-4a01-8c88-7072f09c42eb" (UID: "39ad8b37-a17c-4a01-8c88-7072f09c42eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.929625 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ad8b37-a17c-4a01-8c88-7072f09c42eb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.949960 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rlsd8"] Oct 05 08:32:17 crc kubenswrapper[4935]: I1005 08:32:17.954788 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rlsd8"] Oct 05 08:32:18 crc kubenswrapper[4935]: I1005 08:32:18.786829 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" path="/var/lib/kubelet/pods/39ad8b37-a17c-4a01-8c88-7072f09c42eb/volumes" Oct 05 08:32:48 crc kubenswrapper[4935]: I1005 08:32:48.610772 4935 generic.go:334] "Generic (PLEG): container finished" podID="f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377" containerID="d0ddee54ea9f08be12ab9f6ce977b9c9e03523d87464e394ab6efe2875687f79" exitCode=0 Oct 05 08:32:48 crc kubenswrapper[4935]: I1005 08:32:48.610879 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377","Type":"ContainerDied","Data":"d0ddee54ea9f08be12ab9f6ce977b9c9e03523d87464e394ab6efe2875687f79"} Oct 05 08:32:49 crc kubenswrapper[4935]: I1005 08:32:49.628095 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377","Type":"ContainerStarted","Data":"8ea082e39552e344889c6b2cc756f8f3fbd2ce89b2f74fb59eed70d61bf1a4e2"} Oct 05 08:32:49 crc kubenswrapper[4935]: I1005 08:32:49.628759 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 05 08:32:49 crc kubenswrapper[4935]: I1005 08:32:49.630657 4935 generic.go:334] "Generic (PLEG): container finished" podID="0a21d4cb-88f1-4bec-ad29-97e503d2dad5" containerID="da588cb280e0a39caeb1ec3b876aa1e82fea050b29a1e5b103216f6951acf70d" exitCode=0 Oct 05 08:32:49 crc kubenswrapper[4935]: I1005 08:32:49.630717 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0a21d4cb-88f1-4bec-ad29-97e503d2dad5","Type":"ContainerDied","Data":"da588cb280e0a39caeb1ec3b876aa1e82fea050b29a1e5b103216f6951acf70d"} Oct 05 08:32:49 crc kubenswrapper[4935]: I1005 08:32:49.718765 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.718710608 podStartE2EDuration="36.718710608s" podCreationTimestamp="2025-10-05 08:32:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:32:49.669460828 +0000 UTC m=+6003.552087298" watchObservedRunningTime="2025-10-05 08:32:49.718710608 +0000 UTC m=+6003.601337108" Oct 05 08:32:50 crc kubenswrapper[4935]: I1005 08:32:50.639986 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0a21d4cb-88f1-4bec-ad29-97e503d2dad5","Type":"ContainerStarted","Data":"039371996f858e9a6ec22ffd5f3276a29d24f9fb64512c2b8d5723ba9eb3aebc"} Oct 05 08:32:50 crc kubenswrapper[4935]: I1005 08:32:50.641086 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:32:50 crc kubenswrapper[4935]: I1005 08:32:50.677973 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.677946327 podStartE2EDuration="36.677946327s" podCreationTimestamp="2025-10-05 08:32:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:32:50.677597138 +0000 UTC m=+6004.560223618" watchObservedRunningTime="2025-10-05 08:32:50.677946327 +0000 UTC m=+6004.560572827" Oct 05 08:33:03 crc kubenswrapper[4935]: I1005 08:33:03.712031 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 05 08:33:04 crc kubenswrapper[4935]: I1005 08:33:04.714967 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.043641 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:33:13 crc kubenswrapper[4935]: E1005 08:33:13.044802 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d029017e-6368-4136-b73a-1fac2ff17a1a" containerName="init" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.044825 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d029017e-6368-4136-b73a-1fac2ff17a1a" containerName="init" Oct 05 08:33:13 crc kubenswrapper[4935]: E1005 08:33:13.044860 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d029017e-6368-4136-b73a-1fac2ff17a1a" containerName="dnsmasq-dns" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.044872 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d029017e-6368-4136-b73a-1fac2ff17a1a" containerName="dnsmasq-dns" Oct 05 08:33:13 crc kubenswrapper[4935]: E1005 08:33:13.044919 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerName="registry-server" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.044933 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerName="registry-server" Oct 05 08:33:13 crc kubenswrapper[4935]: E1005 08:33:13.044964 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerName="extract-content" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.044975 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerName="extract-content" Oct 05 08:33:13 crc kubenswrapper[4935]: E1005 08:33:13.044999 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerName="extract-utilities" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.045011 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerName="extract-utilities" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.045351 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ad8b37-a17c-4a01-8c88-7072f09c42eb" containerName="registry-server" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.045399 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d029017e-6368-4136-b73a-1fac2ff17a1a" containerName="dnsmasq-dns" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.046321 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.049741 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sw94x" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.056712 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.118759 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sls74\" (UniqueName: \"kubernetes.io/projected/37b69d33-d143-4c66-8992-035119cf0247-kube-api-access-sls74\") pod \"mariadb-client-1-default\" (UID: \"37b69d33-d143-4c66-8992-035119cf0247\") " pod="openstack/mariadb-client-1-default" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.219767 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sls74\" (UniqueName: \"kubernetes.io/projected/37b69d33-d143-4c66-8992-035119cf0247-kube-api-access-sls74\") pod \"mariadb-client-1-default\" (UID: \"37b69d33-d143-4c66-8992-035119cf0247\") " pod="openstack/mariadb-client-1-default" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.248620 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sls74\" (UniqueName: \"kubernetes.io/projected/37b69d33-d143-4c66-8992-035119cf0247-kube-api-access-sls74\") pod \"mariadb-client-1-default\" (UID: \"37b69d33-d143-4c66-8992-035119cf0247\") " pod="openstack/mariadb-client-1-default" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.375192 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 05 08:33:13 crc kubenswrapper[4935]: I1005 08:33:13.895541 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:33:13 crc kubenswrapper[4935]: W1005 08:33:13.900367 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37b69d33_d143_4c66_8992_035119cf0247.slice/crio-e587ee25c7594175d52cf2e62bb9183c9c778dda817a689e62eb34cd5a577527 WatchSource:0}: Error finding container e587ee25c7594175d52cf2e62bb9183c9c778dda817a689e62eb34cd5a577527: Status 404 returned error can't find the container with id e587ee25c7594175d52cf2e62bb9183c9c778dda817a689e62eb34cd5a577527 Oct 05 08:33:14 crc kubenswrapper[4935]: I1005 08:33:14.855486 4935 generic.go:334] "Generic (PLEG): container finished" podID="37b69d33-d143-4c66-8992-035119cf0247" containerID="efd01fc3540dc655043fbcb0668eaf29f29b7c99eea32399808691c15ffd978d" exitCode=0 Oct 05 08:33:14 crc kubenswrapper[4935]: I1005 08:33:14.855630 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"37b69d33-d143-4c66-8992-035119cf0247","Type":"ContainerDied","Data":"efd01fc3540dc655043fbcb0668eaf29f29b7c99eea32399808691c15ffd978d"} Oct 05 08:33:14 crc kubenswrapper[4935]: I1005 08:33:14.856381 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"37b69d33-d143-4c66-8992-035119cf0247","Type":"ContainerStarted","Data":"e587ee25c7594175d52cf2e62bb9183c9c778dda817a689e62eb34cd5a577527"} Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.275772 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.306712 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_37b69d33-d143-4c66-8992-035119cf0247/mariadb-client-1-default/0.log" Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.337276 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.341678 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.365024 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sls74\" (UniqueName: \"kubernetes.io/projected/37b69d33-d143-4c66-8992-035119cf0247-kube-api-access-sls74\") pod \"37b69d33-d143-4c66-8992-035119cf0247\" (UID: \"37b69d33-d143-4c66-8992-035119cf0247\") " Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.370488 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b69d33-d143-4c66-8992-035119cf0247-kube-api-access-sls74" (OuterVolumeSpecName: "kube-api-access-sls74") pod "37b69d33-d143-4c66-8992-035119cf0247" (UID: "37b69d33-d143-4c66-8992-035119cf0247"). InnerVolumeSpecName "kube-api-access-sls74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.466983 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sls74\" (UniqueName: \"kubernetes.io/projected/37b69d33-d143-4c66-8992-035119cf0247-kube-api-access-sls74\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.787729 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b69d33-d143-4c66-8992-035119cf0247" path="/var/lib/kubelet/pods/37b69d33-d143-4c66-8992-035119cf0247/volumes" Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.873296 4935 scope.go:117] "RemoveContainer" containerID="efd01fc3540dc655043fbcb0668eaf29f29b7c99eea32399808691c15ffd978d" Oct 05 08:33:16 crc kubenswrapper[4935]: I1005 08:33:16.873414 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 05 08:33:17 crc kubenswrapper[4935]: I1005 08:33:17.858690 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:33:17 crc kubenswrapper[4935]: E1005 08:33:17.859325 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b69d33-d143-4c66-8992-035119cf0247" containerName="mariadb-client-1-default" Oct 05 08:33:17 crc kubenswrapper[4935]: I1005 08:33:17.859341 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b69d33-d143-4c66-8992-035119cf0247" containerName="mariadb-client-1-default" Oct 05 08:33:17 crc kubenswrapper[4935]: I1005 08:33:17.859532 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b69d33-d143-4c66-8992-035119cf0247" containerName="mariadb-client-1-default" Oct 05 08:33:17 crc kubenswrapper[4935]: I1005 08:33:17.860112 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 05 08:33:17 crc kubenswrapper[4935]: I1005 08:33:17.863006 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sw94x" Oct 05 08:33:17 crc kubenswrapper[4935]: I1005 08:33:17.873698 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:33:17 crc kubenswrapper[4935]: I1005 08:33:17.888808 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbd8b\" (UniqueName: \"kubernetes.io/projected/c4d78b84-fc01-49d0-8819-871caf80f425-kube-api-access-hbd8b\") pod \"mariadb-client-2-default\" (UID: \"c4d78b84-fc01-49d0-8819-871caf80f425\") " pod="openstack/mariadb-client-2-default" Oct 05 08:33:17 crc kubenswrapper[4935]: I1005 08:33:17.991481 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbd8b\" (UniqueName: \"kubernetes.io/projected/c4d78b84-fc01-49d0-8819-871caf80f425-kube-api-access-hbd8b\") pod \"mariadb-client-2-default\" (UID: \"c4d78b84-fc01-49d0-8819-871caf80f425\") " pod="openstack/mariadb-client-2-default" Oct 05 08:33:18 crc kubenswrapper[4935]: I1005 08:33:18.024646 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbd8b\" (UniqueName: \"kubernetes.io/projected/c4d78b84-fc01-49d0-8819-871caf80f425-kube-api-access-hbd8b\") pod \"mariadb-client-2-default\" (UID: \"c4d78b84-fc01-49d0-8819-871caf80f425\") " pod="openstack/mariadb-client-2-default" Oct 05 08:33:18 crc kubenswrapper[4935]: I1005 08:33:18.181086 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 05 08:33:18 crc kubenswrapper[4935]: I1005 08:33:18.648295 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:33:18 crc kubenswrapper[4935]: I1005 08:33:18.893509 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"c4d78b84-fc01-49d0-8819-871caf80f425","Type":"ContainerStarted","Data":"f5e1b21459361477ef8e4f902d65a86c7f81edd0e8ce39990e07e3d356fbb720"} Oct 05 08:33:18 crc kubenswrapper[4935]: I1005 08:33:18.893570 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"c4d78b84-fc01-49d0-8819-871caf80f425","Type":"ContainerStarted","Data":"02cc567876df3d9e0dcfd3d876d9474cdae6f2449b9979b3b6603b7dc6442091"} Oct 05 08:33:18 crc kubenswrapper[4935]: I1005 08:33:18.909051 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.9090310019999999 podStartE2EDuration="1.909031002s" podCreationTimestamp="2025-10-05 08:33:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:33:18.906859255 +0000 UTC m=+6032.789485725" watchObservedRunningTime="2025-10-05 08:33:18.909031002 +0000 UTC m=+6032.791657462" Oct 05 08:33:19 crc kubenswrapper[4935]: I1005 08:33:19.902814 4935 generic.go:334] "Generic (PLEG): container finished" podID="c4d78b84-fc01-49d0-8819-871caf80f425" containerID="f5e1b21459361477ef8e4f902d65a86c7f81edd0e8ce39990e07e3d356fbb720" exitCode=0 Oct 05 08:33:19 crc kubenswrapper[4935]: I1005 08:33:19.902957 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"c4d78b84-fc01-49d0-8819-871caf80f425","Type":"ContainerDied","Data":"f5e1b21459361477ef8e4f902d65a86c7f81edd0e8ce39990e07e3d356fbb720"} Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.306720 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.342289 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.348158 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.443614 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbd8b\" (UniqueName: \"kubernetes.io/projected/c4d78b84-fc01-49d0-8819-871caf80f425-kube-api-access-hbd8b\") pod \"c4d78b84-fc01-49d0-8819-871caf80f425\" (UID: \"c4d78b84-fc01-49d0-8819-871caf80f425\") " Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.450424 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4d78b84-fc01-49d0-8819-871caf80f425-kube-api-access-hbd8b" (OuterVolumeSpecName: "kube-api-access-hbd8b") pod "c4d78b84-fc01-49d0-8819-871caf80f425" (UID: "c4d78b84-fc01-49d0-8819-871caf80f425"). InnerVolumeSpecName "kube-api-access-hbd8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.546121 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbd8b\" (UniqueName: \"kubernetes.io/projected/c4d78b84-fc01-49d0-8819-871caf80f425-kube-api-access-hbd8b\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.920707 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02cc567876df3d9e0dcfd3d876d9474cdae6f2449b9979b3b6603b7dc6442091" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.920783 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.928036 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:33:21 crc kubenswrapper[4935]: E1005 08:33:21.928566 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d78b84-fc01-49d0-8819-871caf80f425" containerName="mariadb-client-2-default" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.928602 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d78b84-fc01-49d0-8819-871caf80f425" containerName="mariadb-client-2-default" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.928963 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4d78b84-fc01-49d0-8819-871caf80f425" containerName="mariadb-client-2-default" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.929949 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.933647 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sw94x" Oct 05 08:33:21 crc kubenswrapper[4935]: I1005 08:33:21.938866 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.055112 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcd4v\" (UniqueName: \"kubernetes.io/projected/ee8ac79a-23a2-4592-b366-346a689497f9-kube-api-access-rcd4v\") pod \"mariadb-client-1\" (UID: \"ee8ac79a-23a2-4592-b366-346a689497f9\") " pod="openstack/mariadb-client-1" Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.157348 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcd4v\" (UniqueName: \"kubernetes.io/projected/ee8ac79a-23a2-4592-b366-346a689497f9-kube-api-access-rcd4v\") pod \"mariadb-client-1\" (UID: \"ee8ac79a-23a2-4592-b366-346a689497f9\") " pod="openstack/mariadb-client-1" Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.190862 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcd4v\" (UniqueName: \"kubernetes.io/projected/ee8ac79a-23a2-4592-b366-346a689497f9-kube-api-access-rcd4v\") pod \"mariadb-client-1\" (UID: \"ee8ac79a-23a2-4592-b366-346a689497f9\") " pod="openstack/mariadb-client-1" Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.274291 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.668539 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:33:22 crc kubenswrapper[4935]: W1005 08:33:22.674879 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee8ac79a_23a2_4592_b366_346a689497f9.slice/crio-5ce563108a95ff7e2400d2465b76526a26dda08cfc09a21c2d2212be7787e28f WatchSource:0}: Error finding container 5ce563108a95ff7e2400d2465b76526a26dda08cfc09a21c2d2212be7787e28f: Status 404 returned error can't find the container with id 5ce563108a95ff7e2400d2465b76526a26dda08cfc09a21c2d2212be7787e28f Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.792412 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4d78b84-fc01-49d0-8819-871caf80f425" path="/var/lib/kubelet/pods/c4d78b84-fc01-49d0-8819-871caf80f425/volumes" Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.929207 4935 generic.go:334] "Generic (PLEG): container finished" podID="ee8ac79a-23a2-4592-b366-346a689497f9" containerID="7d95dd8c0d9389c3dd84a0663e0ea5d7cf3fca93727382becfa9d9c354d18f81" exitCode=0 Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.929258 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"ee8ac79a-23a2-4592-b366-346a689497f9","Type":"ContainerDied","Data":"7d95dd8c0d9389c3dd84a0663e0ea5d7cf3fca93727382becfa9d9c354d18f81"} Oct 05 08:33:22 crc kubenswrapper[4935]: I1005 08:33:22.929287 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"ee8ac79a-23a2-4592-b366-346a689497f9","Type":"ContainerStarted","Data":"5ce563108a95ff7e2400d2465b76526a26dda08cfc09a21c2d2212be7787e28f"} Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.368276 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.385591 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_ee8ac79a-23a2-4592-b366-346a689497f9/mariadb-client-1/0.log" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.415376 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.420225 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.497350 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcd4v\" (UniqueName: \"kubernetes.io/projected/ee8ac79a-23a2-4592-b366-346a689497f9-kube-api-access-rcd4v\") pod \"ee8ac79a-23a2-4592-b366-346a689497f9\" (UID: \"ee8ac79a-23a2-4592-b366-346a689497f9\") " Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.503055 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee8ac79a-23a2-4592-b366-346a689497f9-kube-api-access-rcd4v" (OuterVolumeSpecName: "kube-api-access-rcd4v") pod "ee8ac79a-23a2-4592-b366-346a689497f9" (UID: "ee8ac79a-23a2-4592-b366-346a689497f9"). InnerVolumeSpecName "kube-api-access-rcd4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.599505 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcd4v\" (UniqueName: \"kubernetes.io/projected/ee8ac79a-23a2-4592-b366-346a689497f9-kube-api-access-rcd4v\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.789209 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee8ac79a-23a2-4592-b366-346a689497f9" path="/var/lib/kubelet/pods/ee8ac79a-23a2-4592-b366-346a689497f9/volumes" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.953198 4935 scope.go:117] "RemoveContainer" containerID="7d95dd8c0d9389c3dd84a0663e0ea5d7cf3fca93727382becfa9d9c354d18f81" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.953318 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.959529 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:33:24 crc kubenswrapper[4935]: E1005 08:33:24.959831 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee8ac79a-23a2-4592-b366-346a689497f9" containerName="mariadb-client-1" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.959847 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee8ac79a-23a2-4592-b366-346a689497f9" containerName="mariadb-client-1" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.960002 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee8ac79a-23a2-4592-b366-346a689497f9" containerName="mariadb-client-1" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.960460 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.971150 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sw94x" Oct 05 08:33:24 crc kubenswrapper[4935]: I1005 08:33:24.986949 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:33:25 crc kubenswrapper[4935]: I1005 08:33:25.107643 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crt9m\" (UniqueName: \"kubernetes.io/projected/465cc37b-a73a-4ca9-a3be-25a6080e5584-kube-api-access-crt9m\") pod \"mariadb-client-4-default\" (UID: \"465cc37b-a73a-4ca9-a3be-25a6080e5584\") " pod="openstack/mariadb-client-4-default" Oct 05 08:33:25 crc kubenswrapper[4935]: I1005 08:33:25.209645 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crt9m\" (UniqueName: \"kubernetes.io/projected/465cc37b-a73a-4ca9-a3be-25a6080e5584-kube-api-access-crt9m\") pod \"mariadb-client-4-default\" (UID: \"465cc37b-a73a-4ca9-a3be-25a6080e5584\") " pod="openstack/mariadb-client-4-default" Oct 05 08:33:25 crc kubenswrapper[4935]: I1005 08:33:25.234638 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crt9m\" (UniqueName: \"kubernetes.io/projected/465cc37b-a73a-4ca9-a3be-25a6080e5584-kube-api-access-crt9m\") pod \"mariadb-client-4-default\" (UID: \"465cc37b-a73a-4ca9-a3be-25a6080e5584\") " pod="openstack/mariadb-client-4-default" Oct 05 08:33:25 crc kubenswrapper[4935]: I1005 08:33:25.331736 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 05 08:33:25 crc kubenswrapper[4935]: I1005 08:33:25.746372 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:33:25 crc kubenswrapper[4935]: I1005 08:33:25.960478 4935 generic.go:334] "Generic (PLEG): container finished" podID="465cc37b-a73a-4ca9-a3be-25a6080e5584" containerID="a10a05dbab6e24197a885af69cbbe6754287ffa4b1429800efbbfca4336b1aab" exitCode=0 Oct 05 08:33:25 crc kubenswrapper[4935]: I1005 08:33:25.960552 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"465cc37b-a73a-4ca9-a3be-25a6080e5584","Type":"ContainerDied","Data":"a10a05dbab6e24197a885af69cbbe6754287ffa4b1429800efbbfca4336b1aab"} Oct 05 08:33:25 crc kubenswrapper[4935]: I1005 08:33:25.960878 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"465cc37b-a73a-4ca9-a3be-25a6080e5584","Type":"ContainerStarted","Data":"2d0e1ef62e63d90045bd6808b7a2417f03bc087efb8a03dc88523783ac9b3d31"} Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.408163 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.428879 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_465cc37b-a73a-4ca9-a3be-25a6080e5584/mariadb-client-4-default/0.log" Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.466720 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.474235 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.549832 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crt9m\" (UniqueName: \"kubernetes.io/projected/465cc37b-a73a-4ca9-a3be-25a6080e5584-kube-api-access-crt9m\") pod \"465cc37b-a73a-4ca9-a3be-25a6080e5584\" (UID: \"465cc37b-a73a-4ca9-a3be-25a6080e5584\") " Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.556945 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/465cc37b-a73a-4ca9-a3be-25a6080e5584-kube-api-access-crt9m" (OuterVolumeSpecName: "kube-api-access-crt9m") pod "465cc37b-a73a-4ca9-a3be-25a6080e5584" (UID: "465cc37b-a73a-4ca9-a3be-25a6080e5584"). InnerVolumeSpecName "kube-api-access-crt9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.651793 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crt9m\" (UniqueName: \"kubernetes.io/projected/465cc37b-a73a-4ca9-a3be-25a6080e5584-kube-api-access-crt9m\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.988933 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d0e1ef62e63d90045bd6808b7a2417f03bc087efb8a03dc88523783ac9b3d31" Oct 05 08:33:27 crc kubenswrapper[4935]: I1005 08:33:27.989029 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 05 08:33:28 crc kubenswrapper[4935]: I1005 08:33:28.791855 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="465cc37b-a73a-4ca9-a3be-25a6080e5584" path="/var/lib/kubelet/pods/465cc37b-a73a-4ca9-a3be-25a6080e5584/volumes" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.392292 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:33:31 crc kubenswrapper[4935]: E1005 08:33:31.393086 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465cc37b-a73a-4ca9-a3be-25a6080e5584" containerName="mariadb-client-4-default" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.393107 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="465cc37b-a73a-4ca9-a3be-25a6080e5584" containerName="mariadb-client-4-default" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.393384 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="465cc37b-a73a-4ca9-a3be-25a6080e5584" containerName="mariadb-client-4-default" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.394174 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.401685 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.427259 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sw94x" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.528354 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wzgd\" (UniqueName: \"kubernetes.io/projected/6e85f93e-0f91-4d35-a1ea-9d80a766d844-kube-api-access-2wzgd\") pod \"mariadb-client-5-default\" (UID: \"6e85f93e-0f91-4d35-a1ea-9d80a766d844\") " pod="openstack/mariadb-client-5-default" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.629805 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wzgd\" (UniqueName: \"kubernetes.io/projected/6e85f93e-0f91-4d35-a1ea-9d80a766d844-kube-api-access-2wzgd\") pod \"mariadb-client-5-default\" (UID: \"6e85f93e-0f91-4d35-a1ea-9d80a766d844\") " pod="openstack/mariadb-client-5-default" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.655776 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wzgd\" (UniqueName: \"kubernetes.io/projected/6e85f93e-0f91-4d35-a1ea-9d80a766d844-kube-api-access-2wzgd\") pod \"mariadb-client-5-default\" (UID: \"6e85f93e-0f91-4d35-a1ea-9d80a766d844\") " pod="openstack/mariadb-client-5-default" Oct 05 08:33:31 crc kubenswrapper[4935]: I1005 08:33:31.748269 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 05 08:33:32 crc kubenswrapper[4935]: I1005 08:33:32.268183 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:33:33 crc kubenswrapper[4935]: I1005 08:33:33.041216 4935 generic.go:334] "Generic (PLEG): container finished" podID="6e85f93e-0f91-4d35-a1ea-9d80a766d844" containerID="21d4509c0d8b7db382846e91b70ad87b43f66a88187ca85d39269c729dc57bb4" exitCode=0 Oct 05 08:33:33 crc kubenswrapper[4935]: I1005 08:33:33.041293 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"6e85f93e-0f91-4d35-a1ea-9d80a766d844","Type":"ContainerDied","Data":"21d4509c0d8b7db382846e91b70ad87b43f66a88187ca85d39269c729dc57bb4"} Oct 05 08:33:33 crc kubenswrapper[4935]: I1005 08:33:33.041582 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"6e85f93e-0f91-4d35-a1ea-9d80a766d844","Type":"ContainerStarted","Data":"a78cf3a703b6cb61171105b6c1042d52e06d83d2caa583f7b13d8d45e330a877"} Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.525507 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.551208 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_6e85f93e-0f91-4d35-a1ea-9d80a766d844/mariadb-client-5-default/0.log" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.583516 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.591043 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.708026 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wzgd\" (UniqueName: \"kubernetes.io/projected/6e85f93e-0f91-4d35-a1ea-9d80a766d844-kube-api-access-2wzgd\") pod \"6e85f93e-0f91-4d35-a1ea-9d80a766d844\" (UID: \"6e85f93e-0f91-4d35-a1ea-9d80a766d844\") " Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.717881 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e85f93e-0f91-4d35-a1ea-9d80a766d844-kube-api-access-2wzgd" (OuterVolumeSpecName: "kube-api-access-2wzgd") pod "6e85f93e-0f91-4d35-a1ea-9d80a766d844" (UID: "6e85f93e-0f91-4d35-a1ea-9d80a766d844"). InnerVolumeSpecName "kube-api-access-2wzgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.752093 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:33:34 crc kubenswrapper[4935]: E1005 08:33:34.752421 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e85f93e-0f91-4d35-a1ea-9d80a766d844" containerName="mariadb-client-5-default" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.752436 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e85f93e-0f91-4d35-a1ea-9d80a766d844" containerName="mariadb-client-5-default" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.752633 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e85f93e-0f91-4d35-a1ea-9d80a766d844" containerName="mariadb-client-5-default" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.753193 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.761978 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.803581 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e85f93e-0f91-4d35-a1ea-9d80a766d844" path="/var/lib/kubelet/pods/6e85f93e-0f91-4d35-a1ea-9d80a766d844/volumes" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.811691 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wzgd\" (UniqueName: \"kubernetes.io/projected/6e85f93e-0f91-4d35-a1ea-9d80a766d844-kube-api-access-2wzgd\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:34 crc kubenswrapper[4935]: I1005 08:33:34.915168 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4hfw\" (UniqueName: \"kubernetes.io/projected/8853b0ab-79e6-41bf-a23c-3a242cd72785-kube-api-access-j4hfw\") pod \"mariadb-client-6-default\" (UID: \"8853b0ab-79e6-41bf-a23c-3a242cd72785\") " pod="openstack/mariadb-client-6-default" Oct 05 08:33:35 crc kubenswrapper[4935]: I1005 08:33:35.016595 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4hfw\" (UniqueName: \"kubernetes.io/projected/8853b0ab-79e6-41bf-a23c-3a242cd72785-kube-api-access-j4hfw\") pod \"mariadb-client-6-default\" (UID: \"8853b0ab-79e6-41bf-a23c-3a242cd72785\") " pod="openstack/mariadb-client-6-default" Oct 05 08:33:35 crc kubenswrapper[4935]: I1005 08:33:35.037800 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4hfw\" (UniqueName: \"kubernetes.io/projected/8853b0ab-79e6-41bf-a23c-3a242cd72785-kube-api-access-j4hfw\") pod \"mariadb-client-6-default\" (UID: \"8853b0ab-79e6-41bf-a23c-3a242cd72785\") " pod="openstack/mariadb-client-6-default" Oct 05 08:33:35 crc kubenswrapper[4935]: I1005 08:33:35.063132 4935 scope.go:117] "RemoveContainer" containerID="21d4509c0d8b7db382846e91b70ad87b43f66a88187ca85d39269c729dc57bb4" Oct 05 08:33:35 crc kubenswrapper[4935]: I1005 08:33:35.063213 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 05 08:33:35 crc kubenswrapper[4935]: I1005 08:33:35.073207 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 05 08:33:35 crc kubenswrapper[4935]: I1005 08:33:35.448220 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:33:35 crc kubenswrapper[4935]: W1005 08:33:35.458220 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8853b0ab_79e6_41bf_a23c_3a242cd72785.slice/crio-5abcbbabcf9f2d2c7a210be1ce67c25d943b510dc4161f41e6ffb8694fa22cae WatchSource:0}: Error finding container 5abcbbabcf9f2d2c7a210be1ce67c25d943b510dc4161f41e6ffb8694fa22cae: Status 404 returned error can't find the container with id 5abcbbabcf9f2d2c7a210be1ce67c25d943b510dc4161f41e6ffb8694fa22cae Oct 05 08:33:36 crc kubenswrapper[4935]: I1005 08:33:36.075179 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"8853b0ab-79e6-41bf-a23c-3a242cd72785","Type":"ContainerStarted","Data":"2f0489c6a5e202aa7be3b66daf469c91d587d6e41883a7dbc0cfd0decb690128"} Oct 05 08:33:36 crc kubenswrapper[4935]: I1005 08:33:36.075272 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"8853b0ab-79e6-41bf-a23c-3a242cd72785","Type":"ContainerStarted","Data":"5abcbbabcf9f2d2c7a210be1ce67c25d943b510dc4161f41e6ffb8694fa22cae"} Oct 05 08:33:36 crc kubenswrapper[4935]: I1005 08:33:36.098695 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=2.098660912 podStartE2EDuration="2.098660912s" podCreationTimestamp="2025-10-05 08:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:33:36.094030899 +0000 UTC m=+6049.976657409" watchObservedRunningTime="2025-10-05 08:33:36.098660912 +0000 UTC m=+6049.981287412" Oct 05 08:33:36 crc kubenswrapper[4935]: I1005 08:33:36.187469 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_8853b0ab-79e6-41bf-a23c-3a242cd72785/mariadb-client-6-default/0.log" Oct 05 08:33:37 crc kubenswrapper[4935]: I1005 08:33:37.088876 4935 generic.go:334] "Generic (PLEG): container finished" podID="8853b0ab-79e6-41bf-a23c-3a242cd72785" containerID="2f0489c6a5e202aa7be3b66daf469c91d587d6e41883a7dbc0cfd0decb690128" exitCode=0 Oct 05 08:33:37 crc kubenswrapper[4935]: I1005 08:33:37.088963 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"8853b0ab-79e6-41bf-a23c-3a242cd72785","Type":"ContainerDied","Data":"2f0489c6a5e202aa7be3b66daf469c91d587d6e41883a7dbc0cfd0decb690128"} Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.581237 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.621383 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.627835 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.777283 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4hfw\" (UniqueName: \"kubernetes.io/projected/8853b0ab-79e6-41bf-a23c-3a242cd72785-kube-api-access-j4hfw\") pod \"8853b0ab-79e6-41bf-a23c-3a242cd72785\" (UID: \"8853b0ab-79e6-41bf-a23c-3a242cd72785\") " Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.784588 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8853b0ab-79e6-41bf-a23c-3a242cd72785-kube-api-access-j4hfw" (OuterVolumeSpecName: "kube-api-access-j4hfw") pod "8853b0ab-79e6-41bf-a23c-3a242cd72785" (UID: "8853b0ab-79e6-41bf-a23c-3a242cd72785"). InnerVolumeSpecName "kube-api-access-j4hfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.820081 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8853b0ab-79e6-41bf-a23c-3a242cd72785" path="/var/lib/kubelet/pods/8853b0ab-79e6-41bf-a23c-3a242cd72785/volumes" Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.824525 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:33:38 crc kubenswrapper[4935]: E1005 08:33:38.825932 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8853b0ab-79e6-41bf-a23c-3a242cd72785" containerName="mariadb-client-6-default" Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.826177 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8853b0ab-79e6-41bf-a23c-3a242cd72785" containerName="mariadb-client-6-default" Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.827127 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8853b0ab-79e6-41bf-a23c-3a242cd72785" containerName="mariadb-client-6-default" Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.829163 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.845988 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.879650 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4hfw\" (UniqueName: \"kubernetes.io/projected/8853b0ab-79e6-41bf-a23c-3a242cd72785-kube-api-access-j4hfw\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:38 crc kubenswrapper[4935]: I1005 08:33:38.981552 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpbtt\" (UniqueName: \"kubernetes.io/projected/44adc603-96af-46b9-a448-690c64afbf72-kube-api-access-cpbtt\") pod \"mariadb-client-7-default\" (UID: \"44adc603-96af-46b9-a448-690c64afbf72\") " pod="openstack/mariadb-client-7-default" Oct 05 08:33:39 crc kubenswrapper[4935]: I1005 08:33:39.084051 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpbtt\" (UniqueName: \"kubernetes.io/projected/44adc603-96af-46b9-a448-690c64afbf72-kube-api-access-cpbtt\") pod \"mariadb-client-7-default\" (UID: \"44adc603-96af-46b9-a448-690c64afbf72\") " pod="openstack/mariadb-client-7-default" Oct 05 08:33:39 crc kubenswrapper[4935]: I1005 08:33:39.101541 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpbtt\" (UniqueName: \"kubernetes.io/projected/44adc603-96af-46b9-a448-690c64afbf72-kube-api-access-cpbtt\") pod \"mariadb-client-7-default\" (UID: \"44adc603-96af-46b9-a448-690c64afbf72\") " pod="openstack/mariadb-client-7-default" Oct 05 08:33:39 crc kubenswrapper[4935]: I1005 08:33:39.114991 4935 scope.go:117] "RemoveContainer" containerID="2f0489c6a5e202aa7be3b66daf469c91d587d6e41883a7dbc0cfd0decb690128" Oct 05 08:33:39 crc kubenswrapper[4935]: I1005 08:33:39.115008 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 05 08:33:39 crc kubenswrapper[4935]: I1005 08:33:39.158723 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 05 08:33:39 crc kubenswrapper[4935]: I1005 08:33:39.543086 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:33:39 crc kubenswrapper[4935]: W1005 08:33:39.551879 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44adc603_96af_46b9_a448_690c64afbf72.slice/crio-a1064243652e806b53472638d0a0726edc8d9d5637b9ca38d3a9c938b2ca8312 WatchSource:0}: Error finding container a1064243652e806b53472638d0a0726edc8d9d5637b9ca38d3a9c938b2ca8312: Status 404 returned error can't find the container with id a1064243652e806b53472638d0a0726edc8d9d5637b9ca38d3a9c938b2ca8312 Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.126953 4935 generic.go:334] "Generic (PLEG): container finished" podID="44adc603-96af-46b9-a448-690c64afbf72" containerID="f1f15072a12c83a10a4e419b8a93dfda5caea678d4c8c14de0f1f5db1fb69fc4" exitCode=0 Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.127069 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"44adc603-96af-46b9-a448-690c64afbf72","Type":"ContainerDied","Data":"f1f15072a12c83a10a4e419b8a93dfda5caea678d4c8c14de0f1f5db1fb69fc4"} Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.128244 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"44adc603-96af-46b9-a448-690c64afbf72","Type":"ContainerStarted","Data":"a1064243652e806b53472638d0a0726edc8d9d5637b9ca38d3a9c938b2ca8312"} Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.738786 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7jnhj"] Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.742780 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.764294 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jnhj"] Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.913509 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fv2x\" (UniqueName: \"kubernetes.io/projected/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-kube-api-access-4fv2x\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.913567 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-utilities\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:40 crc kubenswrapper[4935]: I1005 08:33:40.913625 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-catalog-content\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.014610 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fv2x\" (UniqueName: \"kubernetes.io/projected/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-kube-api-access-4fv2x\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.014676 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-utilities\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.014711 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-catalog-content\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.015385 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-utilities\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.015434 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-catalog-content\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.038029 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fv2x\" (UniqueName: \"kubernetes.io/projected/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-kube-api-access-4fv2x\") pod \"certified-operators-7jnhj\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.076123 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.596882 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.609721 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jnhj"] Oct 05 08:33:41 crc kubenswrapper[4935]: W1005 08:33:41.619243 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d66fb8a_6e44_4572_9eb5_ce7d30aced86.slice/crio-d4299d24ddb17647e6ff08d3c0ae6e9eb684fb7c245397e55f9e407aac533f5d WatchSource:0}: Error finding container d4299d24ddb17647e6ff08d3c0ae6e9eb684fb7c245397e55f9e407aac533f5d: Status 404 returned error can't find the container with id d4299d24ddb17647e6ff08d3c0ae6e9eb684fb7c245397e55f9e407aac533f5d Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.620383 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_44adc603-96af-46b9-a448-690c64afbf72/mariadb-client-7-default/0.log" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.649411 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.654130 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.723132 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpbtt\" (UniqueName: \"kubernetes.io/projected/44adc603-96af-46b9-a448-690c64afbf72-kube-api-access-cpbtt\") pod \"44adc603-96af-46b9-a448-690c64afbf72\" (UID: \"44adc603-96af-46b9-a448-690c64afbf72\") " Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.728500 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44adc603-96af-46b9-a448-690c64afbf72-kube-api-access-cpbtt" (OuterVolumeSpecName: "kube-api-access-cpbtt") pod "44adc603-96af-46b9-a448-690c64afbf72" (UID: "44adc603-96af-46b9-a448-690c64afbf72"). InnerVolumeSpecName "kube-api-access-cpbtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.764159 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:33:41 crc kubenswrapper[4935]: E1005 08:33:41.764565 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44adc603-96af-46b9-a448-690c64afbf72" containerName="mariadb-client-7-default" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.764578 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="44adc603-96af-46b9-a448-690c64afbf72" containerName="mariadb-client-7-default" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.764793 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="44adc603-96af-46b9-a448-690c64afbf72" containerName="mariadb-client-7-default" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.765621 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.769430 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.825988 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rqch\" (UniqueName: \"kubernetes.io/projected/9717f20e-9978-474a-b505-6583a2556951-kube-api-access-2rqch\") pod \"mariadb-client-2\" (UID: \"9717f20e-9978-474a-b505-6583a2556951\") " pod="openstack/mariadb-client-2" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.826076 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpbtt\" (UniqueName: \"kubernetes.io/projected/44adc603-96af-46b9-a448-690c64afbf72-kube-api-access-cpbtt\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.927010 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rqch\" (UniqueName: \"kubernetes.io/projected/9717f20e-9978-474a-b505-6583a2556951-kube-api-access-2rqch\") pod \"mariadb-client-2\" (UID: \"9717f20e-9978-474a-b505-6583a2556951\") " pod="openstack/mariadb-client-2" Oct 05 08:33:41 crc kubenswrapper[4935]: I1005 08:33:41.944462 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rqch\" (UniqueName: \"kubernetes.io/projected/9717f20e-9978-474a-b505-6583a2556951-kube-api-access-2rqch\") pod \"mariadb-client-2\" (UID: \"9717f20e-9978-474a-b505-6583a2556951\") " pod="openstack/mariadb-client-2" Oct 05 08:33:42 crc kubenswrapper[4935]: I1005 08:33:42.085743 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 05 08:33:42 crc kubenswrapper[4935]: I1005 08:33:42.152405 4935 generic.go:334] "Generic (PLEG): container finished" podID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerID="8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd" exitCode=0 Oct 05 08:33:42 crc kubenswrapper[4935]: I1005 08:33:42.152641 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jnhj" event={"ID":"7d66fb8a-6e44-4572-9eb5-ce7d30aced86","Type":"ContainerDied","Data":"8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd"} Oct 05 08:33:42 crc kubenswrapper[4935]: I1005 08:33:42.152819 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jnhj" event={"ID":"7d66fb8a-6e44-4572-9eb5-ce7d30aced86","Type":"ContainerStarted","Data":"d4299d24ddb17647e6ff08d3c0ae6e9eb684fb7c245397e55f9e407aac533f5d"} Oct 05 08:33:42 crc kubenswrapper[4935]: I1005 08:33:42.157227 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1064243652e806b53472638d0a0726edc8d9d5637b9ca38d3a9c938b2ca8312" Oct 05 08:33:42 crc kubenswrapper[4935]: I1005 08:33:42.157261 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 05 08:33:42 crc kubenswrapper[4935]: I1005 08:33:42.586601 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:33:42 crc kubenswrapper[4935]: I1005 08:33:42.786778 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44adc603-96af-46b9-a448-690c64afbf72" path="/var/lib/kubelet/pods/44adc603-96af-46b9-a448-690c64afbf72/volumes" Oct 05 08:33:43 crc kubenswrapper[4935]: I1005 08:33:43.170576 4935 generic.go:334] "Generic (PLEG): container finished" podID="9717f20e-9978-474a-b505-6583a2556951" containerID="a105c67039a038c0d8bb017f3ec95ac2f2a0a2fedc84a6fb5940ef9b165c9c2a" exitCode=0 Oct 05 08:33:43 crc kubenswrapper[4935]: I1005 08:33:43.170681 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"9717f20e-9978-474a-b505-6583a2556951","Type":"ContainerDied","Data":"a105c67039a038c0d8bb017f3ec95ac2f2a0a2fedc84a6fb5940ef9b165c9c2a"} Oct 05 08:33:43 crc kubenswrapper[4935]: I1005 08:33:43.170723 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"9717f20e-9978-474a-b505-6583a2556951","Type":"ContainerStarted","Data":"9c4693a56390580e98ac25f61c6cdd20cf030dd2f54826fe99b4e5c438123a50"} Oct 05 08:33:43 crc kubenswrapper[4935]: I1005 08:33:43.172737 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jnhj" event={"ID":"7d66fb8a-6e44-4572-9eb5-ce7d30aced86","Type":"ContainerStarted","Data":"bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775"} Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.186489 4935 generic.go:334] "Generic (PLEG): container finished" podID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerID="bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775" exitCode=0 Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.186538 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jnhj" event={"ID":"7d66fb8a-6e44-4572-9eb5-ce7d30aced86","Type":"ContainerDied","Data":"bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775"} Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.186930 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jnhj" event={"ID":"7d66fb8a-6e44-4572-9eb5-ce7d30aced86","Type":"ContainerStarted","Data":"7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a"} Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.211522 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7jnhj" podStartSLOduration=2.763346699 podStartE2EDuration="4.211482154s" podCreationTimestamp="2025-10-05 08:33:40 +0000 UTC" firstStartedPulling="2025-10-05 08:33:42.154190793 +0000 UTC m=+6056.036817263" lastFinishedPulling="2025-10-05 08:33:43.602326268 +0000 UTC m=+6057.484952718" observedRunningTime="2025-10-05 08:33:44.209869771 +0000 UTC m=+6058.092496311" watchObservedRunningTime="2025-10-05 08:33:44.211482154 +0000 UTC m=+6058.094108614" Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.289446 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.289510 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.572546 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.588648 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_9717f20e-9978-474a-b505-6583a2556951/mariadb-client-2/0.log" Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.609827 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.616256 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.767860 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rqch\" (UniqueName: \"kubernetes.io/projected/9717f20e-9978-474a-b505-6583a2556951-kube-api-access-2rqch\") pod \"9717f20e-9978-474a-b505-6583a2556951\" (UID: \"9717f20e-9978-474a-b505-6583a2556951\") " Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.776299 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9717f20e-9978-474a-b505-6583a2556951-kube-api-access-2rqch" (OuterVolumeSpecName: "kube-api-access-2rqch") pod "9717f20e-9978-474a-b505-6583a2556951" (UID: "9717f20e-9978-474a-b505-6583a2556951"). InnerVolumeSpecName "kube-api-access-2rqch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.788418 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9717f20e-9978-474a-b505-6583a2556951" path="/var/lib/kubelet/pods/9717f20e-9978-474a-b505-6583a2556951/volumes" Oct 05 08:33:44 crc kubenswrapper[4935]: I1005 08:33:44.869999 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rqch\" (UniqueName: \"kubernetes.io/projected/9717f20e-9978-474a-b505-6583a2556951-kube-api-access-2rqch\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:45 crc kubenswrapper[4935]: I1005 08:33:45.197645 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 05 08:33:45 crc kubenswrapper[4935]: I1005 08:33:45.197678 4935 scope.go:117] "RemoveContainer" containerID="a105c67039a038c0d8bb017f3ec95ac2f2a0a2fedc84a6fb5940ef9b165c9c2a" Oct 05 08:33:51 crc kubenswrapper[4935]: I1005 08:33:51.077410 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:51 crc kubenswrapper[4935]: I1005 08:33:51.078356 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:51 crc kubenswrapper[4935]: I1005 08:33:51.164919 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:51 crc kubenswrapper[4935]: I1005 08:33:51.314584 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:51 crc kubenswrapper[4935]: I1005 08:33:51.413108 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jnhj"] Oct 05 08:33:53 crc kubenswrapper[4935]: I1005 08:33:53.272296 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7jnhj" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerName="registry-server" containerID="cri-o://7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a" gracePeriod=2 Oct 05 08:33:53 crc kubenswrapper[4935]: I1005 08:33:53.841871 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:53 crc kubenswrapper[4935]: I1005 08:33:53.915930 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-catalog-content\") pod \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " Oct 05 08:33:53 crc kubenswrapper[4935]: I1005 08:33:53.916020 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fv2x\" (UniqueName: \"kubernetes.io/projected/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-kube-api-access-4fv2x\") pod \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " Oct 05 08:33:53 crc kubenswrapper[4935]: I1005 08:33:53.916198 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-utilities\") pod \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\" (UID: \"7d66fb8a-6e44-4572-9eb5-ce7d30aced86\") " Oct 05 08:33:53 crc kubenswrapper[4935]: I1005 08:33:53.918288 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-utilities" (OuterVolumeSpecName: "utilities") pod "7d66fb8a-6e44-4572-9eb5-ce7d30aced86" (UID: "7d66fb8a-6e44-4572-9eb5-ce7d30aced86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:33:53 crc kubenswrapper[4935]: I1005 08:33:53.922422 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-kube-api-access-4fv2x" (OuterVolumeSpecName: "kube-api-access-4fv2x") pod "7d66fb8a-6e44-4572-9eb5-ce7d30aced86" (UID: "7d66fb8a-6e44-4572-9eb5-ce7d30aced86"). InnerVolumeSpecName "kube-api-access-4fv2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:33:53 crc kubenswrapper[4935]: I1005 08:33:53.975150 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d66fb8a-6e44-4572-9eb5-ce7d30aced86" (UID: "7d66fb8a-6e44-4572-9eb5-ce7d30aced86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.019106 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.019158 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.019182 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fv2x\" (UniqueName: \"kubernetes.io/projected/7d66fb8a-6e44-4572-9eb5-ce7d30aced86-kube-api-access-4fv2x\") on node \"crc\" DevicePath \"\"" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.284383 4935 generic.go:334] "Generic (PLEG): container finished" podID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerID="7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a" exitCode=0 Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.284428 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jnhj" event={"ID":"7d66fb8a-6e44-4572-9eb5-ce7d30aced86","Type":"ContainerDied","Data":"7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a"} Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.284457 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jnhj" event={"ID":"7d66fb8a-6e44-4572-9eb5-ce7d30aced86","Type":"ContainerDied","Data":"d4299d24ddb17647e6ff08d3c0ae6e9eb684fb7c245397e55f9e407aac533f5d"} Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.284477 4935 scope.go:117] "RemoveContainer" containerID="7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.284592 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jnhj" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.322140 4935 scope.go:117] "RemoveContainer" containerID="bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.348386 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jnhj"] Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.351524 4935 scope.go:117] "RemoveContainer" containerID="8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.357957 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7jnhj"] Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.381476 4935 scope.go:117] "RemoveContainer" containerID="7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a" Oct 05 08:33:54 crc kubenswrapper[4935]: E1005 08:33:54.382109 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a\": container with ID starting with 7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a not found: ID does not exist" containerID="7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.382148 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a"} err="failed to get container status \"7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a\": rpc error: code = NotFound desc = could not find container \"7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a\": container with ID starting with 7cb1b961623f9c4b1f042b26f29c6117200c45723f698edc1bfad920484abe7a not found: ID does not exist" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.382174 4935 scope.go:117] "RemoveContainer" containerID="bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775" Oct 05 08:33:54 crc kubenswrapper[4935]: E1005 08:33:54.382534 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775\": container with ID starting with bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775 not found: ID does not exist" containerID="bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.382557 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775"} err="failed to get container status \"bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775\": rpc error: code = NotFound desc = could not find container \"bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775\": container with ID starting with bc1dcd43982f297a0f1e021c628cea3e927f0b9d4075372c23b518cbe8225775 not found: ID does not exist" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.382573 4935 scope.go:117] "RemoveContainer" containerID="8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd" Oct 05 08:33:54 crc kubenswrapper[4935]: E1005 08:33:54.382943 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd\": container with ID starting with 8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd not found: ID does not exist" containerID="8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.383001 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd"} err="failed to get container status \"8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd\": rpc error: code = NotFound desc = could not find container \"8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd\": container with ID starting with 8dddfc64404324243891b2bf937287c6ef80d7cd7c3ff7b95ce5bd76450979bd not found: ID does not exist" Oct 05 08:33:54 crc kubenswrapper[4935]: I1005 08:33:54.790684 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" path="/var/lib/kubelet/pods/7d66fb8a-6e44-4572-9eb5-ce7d30aced86/volumes" Oct 05 08:34:14 crc kubenswrapper[4935]: I1005 08:34:14.289635 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:34:14 crc kubenswrapper[4935]: I1005 08:34:14.290402 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.289108 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.289800 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.289858 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.290582 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c98eb32c261a6373bd85357643e260ea4fb13065049a0fdbd29241582949a384"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.290640 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://c98eb32c261a6373bd85357643e260ea4fb13065049a0fdbd29241582949a384" gracePeriod=600 Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.772085 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="c98eb32c261a6373bd85357643e260ea4fb13065049a0fdbd29241582949a384" exitCode=0 Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.772253 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"c98eb32c261a6373bd85357643e260ea4fb13065049a0fdbd29241582949a384"} Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.772501 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b"} Oct 05 08:34:44 crc kubenswrapper[4935]: I1005 08:34:44.772537 4935 scope.go:117] "RemoveContainer" containerID="8c5228f33c9b816317167b5bde9311ff1c478aece1e4b4bcf0ff5a4debf3c484" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.802004 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nmmx2"] Oct 05 08:34:45 crc kubenswrapper[4935]: E1005 08:34:45.802857 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerName="extract-utilities" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.802876 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerName="extract-utilities" Oct 05 08:34:45 crc kubenswrapper[4935]: E1005 08:34:45.802928 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerName="registry-server" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.802941 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerName="registry-server" Oct 05 08:34:45 crc kubenswrapper[4935]: E1005 08:34:45.802972 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerName="extract-content" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.802984 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerName="extract-content" Oct 05 08:34:45 crc kubenswrapper[4935]: E1005 08:34:45.803010 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9717f20e-9978-474a-b505-6583a2556951" containerName="mariadb-client-2" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.803021 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9717f20e-9978-474a-b505-6583a2556951" containerName="mariadb-client-2" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.803259 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9717f20e-9978-474a-b505-6583a2556951" containerName="mariadb-client-2" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.803294 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d66fb8a-6e44-4572-9eb5-ce7d30aced86" containerName="registry-server" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.805043 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.826039 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nmmx2"] Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.983385 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-catalog-content\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.983725 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-utilities\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:45 crc kubenswrapper[4935]: I1005 08:34:45.983831 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcwqx\" (UniqueName: \"kubernetes.io/projected/c8fbe9cf-2e2c-488d-b64e-b90850924078-kube-api-access-gcwqx\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.085264 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-utilities\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.085333 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcwqx\" (UniqueName: \"kubernetes.io/projected/c8fbe9cf-2e2c-488d-b64e-b90850924078-kube-api-access-gcwqx\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.085368 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-catalog-content\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.085831 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-utilities\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.085912 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-catalog-content\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.112180 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcwqx\" (UniqueName: \"kubernetes.io/projected/c8fbe9cf-2e2c-488d-b64e-b90850924078-kube-api-access-gcwqx\") pod \"community-operators-nmmx2\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.173494 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.658268 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nmmx2"] Oct 05 08:34:46 crc kubenswrapper[4935]: W1005 08:34:46.667297 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8fbe9cf_2e2c_488d_b64e_b90850924078.slice/crio-22313f2d2c1a91de314353b59f2807b0e6544a0eab075046732526397456fd38 WatchSource:0}: Error finding container 22313f2d2c1a91de314353b59f2807b0e6544a0eab075046732526397456fd38: Status 404 returned error can't find the container with id 22313f2d2c1a91de314353b59f2807b0e6544a0eab075046732526397456fd38 Oct 05 08:34:46 crc kubenswrapper[4935]: I1005 08:34:46.808377 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmmx2" event={"ID":"c8fbe9cf-2e2c-488d-b64e-b90850924078","Type":"ContainerStarted","Data":"22313f2d2c1a91de314353b59f2807b0e6544a0eab075046732526397456fd38"} Oct 05 08:34:47 crc kubenswrapper[4935]: I1005 08:34:47.832337 4935 generic.go:334] "Generic (PLEG): container finished" podID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerID="04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb" exitCode=0 Oct 05 08:34:47 crc kubenswrapper[4935]: I1005 08:34:47.832430 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmmx2" event={"ID":"c8fbe9cf-2e2c-488d-b64e-b90850924078","Type":"ContainerDied","Data":"04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb"} Oct 05 08:34:49 crc kubenswrapper[4935]: I1005 08:34:49.855058 4935 generic.go:334] "Generic (PLEG): container finished" podID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerID="d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d" exitCode=0 Oct 05 08:34:49 crc kubenswrapper[4935]: I1005 08:34:49.855112 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmmx2" event={"ID":"c8fbe9cf-2e2c-488d-b64e-b90850924078","Type":"ContainerDied","Data":"d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d"} Oct 05 08:34:50 crc kubenswrapper[4935]: I1005 08:34:50.866681 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmmx2" event={"ID":"c8fbe9cf-2e2c-488d-b64e-b90850924078","Type":"ContainerStarted","Data":"ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61"} Oct 05 08:34:50 crc kubenswrapper[4935]: I1005 08:34:50.897864 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nmmx2" podStartSLOduration=3.480750791 podStartE2EDuration="5.897833166s" podCreationTimestamp="2025-10-05 08:34:45 +0000 UTC" firstStartedPulling="2025-10-05 08:34:47.835421424 +0000 UTC m=+6121.718047914" lastFinishedPulling="2025-10-05 08:34:50.252503789 +0000 UTC m=+6124.135130289" observedRunningTime="2025-10-05 08:34:50.891200299 +0000 UTC m=+6124.773826829" watchObservedRunningTime="2025-10-05 08:34:50.897833166 +0000 UTC m=+6124.780459656" Oct 05 08:34:56 crc kubenswrapper[4935]: I1005 08:34:56.174380 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:56 crc kubenswrapper[4935]: I1005 08:34:56.174775 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:56 crc kubenswrapper[4935]: I1005 08:34:56.224962 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:56 crc kubenswrapper[4935]: I1005 08:34:56.715157 4935 scope.go:117] "RemoveContainer" containerID="8af944ec27a78ab87289bc3295023b1f25f79ee198bedd758226ec4a7b425ef8" Oct 05 08:34:56 crc kubenswrapper[4935]: I1005 08:34:56.984719 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:57 crc kubenswrapper[4935]: I1005 08:34:57.040674 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nmmx2"] Oct 05 08:34:58 crc kubenswrapper[4935]: I1005 08:34:58.949083 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nmmx2" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerName="registry-server" containerID="cri-o://ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61" gracePeriod=2 Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.417133 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.607869 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-utilities\") pod \"c8fbe9cf-2e2c-488d-b64e-b90850924078\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.608187 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-catalog-content\") pod \"c8fbe9cf-2e2c-488d-b64e-b90850924078\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.608233 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcwqx\" (UniqueName: \"kubernetes.io/projected/c8fbe9cf-2e2c-488d-b64e-b90850924078-kube-api-access-gcwqx\") pod \"c8fbe9cf-2e2c-488d-b64e-b90850924078\" (UID: \"c8fbe9cf-2e2c-488d-b64e-b90850924078\") " Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.608629 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-utilities" (OuterVolumeSpecName: "utilities") pod "c8fbe9cf-2e2c-488d-b64e-b90850924078" (UID: "c8fbe9cf-2e2c-488d-b64e-b90850924078"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.615033 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8fbe9cf-2e2c-488d-b64e-b90850924078-kube-api-access-gcwqx" (OuterVolumeSpecName: "kube-api-access-gcwqx") pod "c8fbe9cf-2e2c-488d-b64e-b90850924078" (UID: "c8fbe9cf-2e2c-488d-b64e-b90850924078"). InnerVolumeSpecName "kube-api-access-gcwqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.685414 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8fbe9cf-2e2c-488d-b64e-b90850924078" (UID: "c8fbe9cf-2e2c-488d-b64e-b90850924078"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.709872 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.709972 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcwqx\" (UniqueName: \"kubernetes.io/projected/c8fbe9cf-2e2c-488d-b64e-b90850924078-kube-api-access-gcwqx\") on node \"crc\" DevicePath \"\"" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.709985 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8fbe9cf-2e2c-488d-b64e-b90850924078-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.958551 4935 generic.go:334] "Generic (PLEG): container finished" podID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerID="ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61" exitCode=0 Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.958601 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmmx2" event={"ID":"c8fbe9cf-2e2c-488d-b64e-b90850924078","Type":"ContainerDied","Data":"ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61"} Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.958653 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmmx2" event={"ID":"c8fbe9cf-2e2c-488d-b64e-b90850924078","Type":"ContainerDied","Data":"22313f2d2c1a91de314353b59f2807b0e6544a0eab075046732526397456fd38"} Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.958670 4935 scope.go:117] "RemoveContainer" containerID="ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.958785 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmmx2" Oct 05 08:34:59 crc kubenswrapper[4935]: I1005 08:34:59.980274 4935 scope.go:117] "RemoveContainer" containerID="d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d" Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.001185 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nmmx2"] Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.001567 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nmmx2"] Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.022586 4935 scope.go:117] "RemoveContainer" containerID="04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb" Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.055063 4935 scope.go:117] "RemoveContainer" containerID="ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61" Oct 05 08:35:00 crc kubenswrapper[4935]: E1005 08:35:00.055444 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61\": container with ID starting with ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61 not found: ID does not exist" containerID="ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61" Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.055490 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61"} err="failed to get container status \"ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61\": rpc error: code = NotFound desc = could not find container \"ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61\": container with ID starting with ef286c645109b555954a11a65357c494fdcb3824d3bb42d5805cc87d893b8c61 not found: ID does not exist" Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.055516 4935 scope.go:117] "RemoveContainer" containerID="d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d" Oct 05 08:35:00 crc kubenswrapper[4935]: E1005 08:35:00.055833 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d\": container with ID starting with d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d not found: ID does not exist" containerID="d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d" Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.055881 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d"} err="failed to get container status \"d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d\": rpc error: code = NotFound desc = could not find container \"d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d\": container with ID starting with d8567009178063bf642c0f58de11c0694272dbb1f89affe67b160bbd1a39d15d not found: ID does not exist" Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.055925 4935 scope.go:117] "RemoveContainer" containerID="04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb" Oct 05 08:35:00 crc kubenswrapper[4935]: E1005 08:35:00.056380 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb\": container with ID starting with 04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb not found: ID does not exist" containerID="04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb" Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.056418 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb"} err="failed to get container status \"04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb\": rpc error: code = NotFound desc = could not find container \"04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb\": container with ID starting with 04b7da34762408cb72ae7a4476ed372059dc1513831126866dec890f6c6ceddb not found: ID does not exist" Oct 05 08:35:00 crc kubenswrapper[4935]: I1005 08:35:00.793446 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" path="/var/lib/kubelet/pods/c8fbe9cf-2e2c-488d-b64e-b90850924078/volumes" Oct 05 08:36:44 crc kubenswrapper[4935]: I1005 08:36:44.289753 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:36:44 crc kubenswrapper[4935]: I1005 08:36:44.290815 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.645371 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8lvxx"] Oct 05 08:36:54 crc kubenswrapper[4935]: E1005 08:36:54.646650 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerName="extract-utilities" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.646681 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerName="extract-utilities" Oct 05 08:36:54 crc kubenswrapper[4935]: E1005 08:36:54.646762 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerName="registry-server" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.646782 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerName="registry-server" Oct 05 08:36:54 crc kubenswrapper[4935]: E1005 08:36:54.646818 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerName="extract-content" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.646836 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerName="extract-content" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.647343 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8fbe9cf-2e2c-488d-b64e-b90850924078" containerName="registry-server" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.650251 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.661669 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lvxx"] Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.796461 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-utilities\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.796543 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-catalog-content\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.796611 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg542\" (UniqueName: \"kubernetes.io/projected/81f90944-3612-4abf-8335-cff5b227103c-kube-api-access-kg542\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.898145 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-utilities\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.898238 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-catalog-content\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.898267 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg542\" (UniqueName: \"kubernetes.io/projected/81f90944-3612-4abf-8335-cff5b227103c-kube-api-access-kg542\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.898735 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-utilities\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.899162 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-catalog-content\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.916464 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg542\" (UniqueName: \"kubernetes.io/projected/81f90944-3612-4abf-8335-cff5b227103c-kube-api-access-kg542\") pod \"redhat-marketplace-8lvxx\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:54 crc kubenswrapper[4935]: I1005 08:36:54.996527 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:36:55 crc kubenswrapper[4935]: I1005 08:36:55.412562 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lvxx"] Oct 05 08:36:56 crc kubenswrapper[4935]: I1005 08:36:56.068249 4935 generic.go:334] "Generic (PLEG): container finished" podID="81f90944-3612-4abf-8335-cff5b227103c" containerID="4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80" exitCode=0 Oct 05 08:36:56 crc kubenswrapper[4935]: I1005 08:36:56.068353 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lvxx" event={"ID":"81f90944-3612-4abf-8335-cff5b227103c","Type":"ContainerDied","Data":"4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80"} Oct 05 08:36:56 crc kubenswrapper[4935]: I1005 08:36:56.068510 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lvxx" event={"ID":"81f90944-3612-4abf-8335-cff5b227103c","Type":"ContainerStarted","Data":"1b1f6254e1869ecaf65c9a8a3aee3e4fb3c266dc898c463fd78da616e2c8ef19"} Oct 05 08:36:58 crc kubenswrapper[4935]: I1005 08:36:58.088515 4935 generic.go:334] "Generic (PLEG): container finished" podID="81f90944-3612-4abf-8335-cff5b227103c" containerID="4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418" exitCode=0 Oct 05 08:36:58 crc kubenswrapper[4935]: I1005 08:36:58.088580 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lvxx" event={"ID":"81f90944-3612-4abf-8335-cff5b227103c","Type":"ContainerDied","Data":"4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418"} Oct 05 08:36:59 crc kubenswrapper[4935]: I1005 08:36:59.107586 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lvxx" event={"ID":"81f90944-3612-4abf-8335-cff5b227103c","Type":"ContainerStarted","Data":"50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301"} Oct 05 08:36:59 crc kubenswrapper[4935]: I1005 08:36:59.130687 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8lvxx" podStartSLOduration=2.601126785 podStartE2EDuration="5.130658471s" podCreationTimestamp="2025-10-05 08:36:54 +0000 UTC" firstStartedPulling="2025-10-05 08:36:56.0705692 +0000 UTC m=+6249.953195670" lastFinishedPulling="2025-10-05 08:36:58.600100856 +0000 UTC m=+6252.482727356" observedRunningTime="2025-10-05 08:36:59.126095389 +0000 UTC m=+6253.008721889" watchObservedRunningTime="2025-10-05 08:36:59.130658471 +0000 UTC m=+6253.013284941" Oct 05 08:37:04 crc kubenswrapper[4935]: I1005 08:37:04.997021 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:37:04 crc kubenswrapper[4935]: I1005 08:37:04.997973 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:37:05 crc kubenswrapper[4935]: I1005 08:37:05.049539 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:37:05 crc kubenswrapper[4935]: I1005 08:37:05.256011 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:37:05 crc kubenswrapper[4935]: I1005 08:37:05.317316 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lvxx"] Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.202423 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8lvxx" podUID="81f90944-3612-4abf-8335-cff5b227103c" containerName="registry-server" containerID="cri-o://50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301" gracePeriod=2 Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.655270 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.729778 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg542\" (UniqueName: \"kubernetes.io/projected/81f90944-3612-4abf-8335-cff5b227103c-kube-api-access-kg542\") pod \"81f90944-3612-4abf-8335-cff5b227103c\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.729933 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-utilities\") pod \"81f90944-3612-4abf-8335-cff5b227103c\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.730030 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-catalog-content\") pod \"81f90944-3612-4abf-8335-cff5b227103c\" (UID: \"81f90944-3612-4abf-8335-cff5b227103c\") " Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.731395 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-utilities" (OuterVolumeSpecName: "utilities") pod "81f90944-3612-4abf-8335-cff5b227103c" (UID: "81f90944-3612-4abf-8335-cff5b227103c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.738607 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f90944-3612-4abf-8335-cff5b227103c-kube-api-access-kg542" (OuterVolumeSpecName: "kube-api-access-kg542") pod "81f90944-3612-4abf-8335-cff5b227103c" (UID: "81f90944-3612-4abf-8335-cff5b227103c"). InnerVolumeSpecName "kube-api-access-kg542". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.743225 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81f90944-3612-4abf-8335-cff5b227103c" (UID: "81f90944-3612-4abf-8335-cff5b227103c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.831504 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.831535 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f90944-3612-4abf-8335-cff5b227103c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:07 crc kubenswrapper[4935]: I1005 08:37:07.831548 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg542\" (UniqueName: \"kubernetes.io/projected/81f90944-3612-4abf-8335-cff5b227103c-kube-api-access-kg542\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.215481 4935 generic.go:334] "Generic (PLEG): container finished" podID="81f90944-3612-4abf-8335-cff5b227103c" containerID="50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301" exitCode=0 Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.215561 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lvxx" event={"ID":"81f90944-3612-4abf-8335-cff5b227103c","Type":"ContainerDied","Data":"50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301"} Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.216005 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lvxx" event={"ID":"81f90944-3612-4abf-8335-cff5b227103c","Type":"ContainerDied","Data":"1b1f6254e1869ecaf65c9a8a3aee3e4fb3c266dc898c463fd78da616e2c8ef19"} Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.216042 4935 scope.go:117] "RemoveContainer" containerID="50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.215586 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lvxx" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.242313 4935 scope.go:117] "RemoveContainer" containerID="4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.262404 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lvxx"] Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.270116 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lvxx"] Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.291932 4935 scope.go:117] "RemoveContainer" containerID="4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.320155 4935 scope.go:117] "RemoveContainer" containerID="50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301" Oct 05 08:37:08 crc kubenswrapper[4935]: E1005 08:37:08.320593 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301\": container with ID starting with 50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301 not found: ID does not exist" containerID="50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.320644 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301"} err="failed to get container status \"50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301\": rpc error: code = NotFound desc = could not find container \"50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301\": container with ID starting with 50185e8e5edaf37440d83aaa9f16901ac919269e6ae9cfbc6ac3b444243de301 not found: ID does not exist" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.320677 4935 scope.go:117] "RemoveContainer" containerID="4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418" Oct 05 08:37:08 crc kubenswrapper[4935]: E1005 08:37:08.321212 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418\": container with ID starting with 4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418 not found: ID does not exist" containerID="4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.321239 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418"} err="failed to get container status \"4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418\": rpc error: code = NotFound desc = could not find container \"4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418\": container with ID starting with 4213d973352f6e1a0f222e3b768058abaa14767586384175f8b98dd37384e418 not found: ID does not exist" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.321259 4935 scope.go:117] "RemoveContainer" containerID="4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80" Oct 05 08:37:08 crc kubenswrapper[4935]: E1005 08:37:08.321582 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80\": container with ID starting with 4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80 not found: ID does not exist" containerID="4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.321639 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80"} err="failed to get container status \"4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80\": rpc error: code = NotFound desc = could not find container \"4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80\": container with ID starting with 4c56e9bc69333468ad610912cd90737b0d066d8273d7edd64599b99394334b80 not found: ID does not exist" Oct 05 08:37:08 crc kubenswrapper[4935]: I1005 08:37:08.791576 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f90944-3612-4abf-8335-cff5b227103c" path="/var/lib/kubelet/pods/81f90944-3612-4abf-8335-cff5b227103c/volumes" Oct 05 08:37:14 crc kubenswrapper[4935]: I1005 08:37:14.289995 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:37:14 crc kubenswrapper[4935]: I1005 08:37:14.290486 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.291244 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.291860 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.291967 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.293061 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.293183 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" gracePeriod=600 Oct 05 08:37:44 crc kubenswrapper[4935]: E1005 08:37:44.431795 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.572382 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" exitCode=0 Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.572453 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b"} Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.572537 4935 scope.go:117] "RemoveContainer" containerID="c98eb32c261a6373bd85357643e260ea4fb13065049a0fdbd29241582949a384" Oct 05 08:37:44 crc kubenswrapper[4935]: I1005 08:37:44.572996 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:37:44 crc kubenswrapper[4935]: E1005 08:37:44.573216 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:37:58 crc kubenswrapper[4935]: I1005 08:37:58.778022 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:37:58 crc kubenswrapper[4935]: E1005 08:37:58.779470 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:38:10 crc kubenswrapper[4935]: I1005 08:38:10.777169 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:38:10 crc kubenswrapper[4935]: E1005 08:38:10.777954 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:38:23 crc kubenswrapper[4935]: I1005 08:38:23.777731 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:38:23 crc kubenswrapper[4935]: E1005 08:38:23.778942 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:38:38 crc kubenswrapper[4935]: I1005 08:38:38.778202 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:38:38 crc kubenswrapper[4935]: E1005 08:38:38.779551 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:38:50 crc kubenswrapper[4935]: I1005 08:38:50.778524 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:38:50 crc kubenswrapper[4935]: E1005 08:38:50.779747 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:39:04 crc kubenswrapper[4935]: I1005 08:39:04.777995 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:39:04 crc kubenswrapper[4935]: E1005 08:39:04.779273 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:39:18 crc kubenswrapper[4935]: I1005 08:39:18.777612 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:39:18 crc kubenswrapper[4935]: E1005 08:39:18.778582 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:39:30 crc kubenswrapper[4935]: I1005 08:39:30.778297 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:39:30 crc kubenswrapper[4935]: E1005 08:39:30.778945 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:39:42 crc kubenswrapper[4935]: I1005 08:39:42.778025 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:39:42 crc kubenswrapper[4935]: E1005 08:39:42.778815 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:39:55 crc kubenswrapper[4935]: I1005 08:39:55.777810 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:39:55 crc kubenswrapper[4935]: E1005 08:39:55.779068 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:39:56 crc kubenswrapper[4935]: I1005 08:39:56.884496 4935 scope.go:117] "RemoveContainer" containerID="f1f15072a12c83a10a4e419b8a93dfda5caea678d4c8c14de0f1f5db1fb69fc4" Oct 05 08:39:56 crc kubenswrapper[4935]: I1005 08:39:56.912418 4935 scope.go:117] "RemoveContainer" containerID="f5e1b21459361477ef8e4f902d65a86c7f81edd0e8ce39990e07e3d356fbb720" Oct 05 08:39:56 crc kubenswrapper[4935]: I1005 08:39:56.952864 4935 scope.go:117] "RemoveContainer" containerID="a10a05dbab6e24197a885af69cbbe6754287ffa4b1429800efbbfca4336b1aab" Oct 05 08:40:07 crc kubenswrapper[4935]: I1005 08:40:07.777050 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:40:07 crc kubenswrapper[4935]: E1005 08:40:07.777866 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:40:19 crc kubenswrapper[4935]: I1005 08:40:19.777771 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:40:19 crc kubenswrapper[4935]: E1005 08:40:19.778533 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:40:31 crc kubenswrapper[4935]: I1005 08:40:31.779842 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:40:31 crc kubenswrapper[4935]: E1005 08:40:31.781211 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:40:44 crc kubenswrapper[4935]: I1005 08:40:44.777262 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:40:44 crc kubenswrapper[4935]: E1005 08:40:44.778326 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:40:56 crc kubenswrapper[4935]: I1005 08:40:56.788217 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:40:56 crc kubenswrapper[4935]: E1005 08:40:56.789315 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:41:09 crc kubenswrapper[4935]: I1005 08:41:09.777403 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:41:09 crc kubenswrapper[4935]: E1005 08:41:09.780178 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:41:20 crc kubenswrapper[4935]: I1005 08:41:20.777024 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:41:20 crc kubenswrapper[4935]: E1005 08:41:20.777719 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:41:35 crc kubenswrapper[4935]: I1005 08:41:35.777682 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:41:35 crc kubenswrapper[4935]: E1005 08:41:35.779118 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:41:46 crc kubenswrapper[4935]: I1005 08:41:46.786275 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:41:46 crc kubenswrapper[4935]: E1005 08:41:46.787428 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:42:00 crc kubenswrapper[4935]: I1005 08:42:00.777073 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:42:00 crc kubenswrapper[4935]: E1005 08:42:00.778152 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:42:12 crc kubenswrapper[4935]: I1005 08:42:12.777817 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:42:12 crc kubenswrapper[4935]: E1005 08:42:12.778730 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:42:23 crc kubenswrapper[4935]: I1005 08:42:23.777545 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:42:23 crc kubenswrapper[4935]: E1005 08:42:23.778638 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:42:37 crc kubenswrapper[4935]: I1005 08:42:37.777218 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:42:37 crc kubenswrapper[4935]: E1005 08:42:37.778849 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.642577 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-njgdw"] Oct 05 08:42:42 crc kubenswrapper[4935]: E1005 08:42:42.643714 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f90944-3612-4abf-8335-cff5b227103c" containerName="extract-utilities" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.643737 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f90944-3612-4abf-8335-cff5b227103c" containerName="extract-utilities" Oct 05 08:42:42 crc kubenswrapper[4935]: E1005 08:42:42.643759 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f90944-3612-4abf-8335-cff5b227103c" containerName="extract-content" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.643771 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f90944-3612-4abf-8335-cff5b227103c" containerName="extract-content" Oct 05 08:42:42 crc kubenswrapper[4935]: E1005 08:42:42.643803 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f90944-3612-4abf-8335-cff5b227103c" containerName="registry-server" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.643816 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f90944-3612-4abf-8335-cff5b227103c" containerName="registry-server" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.644187 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f90944-3612-4abf-8335-cff5b227103c" containerName="registry-server" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.646118 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.672476 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-catalog-content\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.672679 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2pf8\" (UniqueName: \"kubernetes.io/projected/44d3b7ce-403d-437b-9c48-d44913198dc7-kube-api-access-k2pf8\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.673107 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-utilities\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.709711 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-njgdw"] Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.775003 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-catalog-content\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.775120 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2pf8\" (UniqueName: \"kubernetes.io/projected/44d3b7ce-403d-437b-9c48-d44913198dc7-kube-api-access-k2pf8\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.775275 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-utilities\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.775762 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-catalog-content\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.775774 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-utilities\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.802530 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2pf8\" (UniqueName: \"kubernetes.io/projected/44d3b7ce-403d-437b-9c48-d44913198dc7-kube-api-access-k2pf8\") pod \"redhat-operators-njgdw\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:42 crc kubenswrapper[4935]: I1005 08:42:42.983316 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:43 crc kubenswrapper[4935]: I1005 08:42:43.257821 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-njgdw"] Oct 05 08:42:43 crc kubenswrapper[4935]: I1005 08:42:43.404622 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njgdw" event={"ID":"44d3b7ce-403d-437b-9c48-d44913198dc7","Type":"ContainerStarted","Data":"2843cf977f2cbb6cdd23b947e8bb2fb66ff5484329dba528d497a0833f9fcb69"} Oct 05 08:42:44 crc kubenswrapper[4935]: I1005 08:42:44.417768 4935 generic.go:334] "Generic (PLEG): container finished" podID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerID="a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752" exitCode=0 Oct 05 08:42:44 crc kubenswrapper[4935]: I1005 08:42:44.417821 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njgdw" event={"ID":"44d3b7ce-403d-437b-9c48-d44913198dc7","Type":"ContainerDied","Data":"a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752"} Oct 05 08:42:44 crc kubenswrapper[4935]: I1005 08:42:44.420294 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:42:46 crc kubenswrapper[4935]: I1005 08:42:46.436881 4935 generic.go:334] "Generic (PLEG): container finished" podID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerID="8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4" exitCode=0 Oct 05 08:42:46 crc kubenswrapper[4935]: I1005 08:42:46.436948 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njgdw" event={"ID":"44d3b7ce-403d-437b-9c48-d44913198dc7","Type":"ContainerDied","Data":"8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4"} Oct 05 08:42:47 crc kubenswrapper[4935]: I1005 08:42:47.449121 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njgdw" event={"ID":"44d3b7ce-403d-437b-9c48-d44913198dc7","Type":"ContainerStarted","Data":"bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4"} Oct 05 08:42:47 crc kubenswrapper[4935]: I1005 08:42:47.469033 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-njgdw" podStartSLOduration=3.022417635 podStartE2EDuration="5.469012159s" podCreationTimestamp="2025-10-05 08:42:42 +0000 UTC" firstStartedPulling="2025-10-05 08:42:44.42003061 +0000 UTC m=+6598.302657070" lastFinishedPulling="2025-10-05 08:42:46.866625104 +0000 UTC m=+6600.749251594" observedRunningTime="2025-10-05 08:42:47.464871708 +0000 UTC m=+6601.347498188" watchObservedRunningTime="2025-10-05 08:42:47.469012159 +0000 UTC m=+6601.351638619" Oct 05 08:42:52 crc kubenswrapper[4935]: I1005 08:42:52.777726 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:42:52 crc kubenswrapper[4935]: I1005 08:42:52.984227 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:52 crc kubenswrapper[4935]: I1005 08:42:52.984293 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:53 crc kubenswrapper[4935]: I1005 08:42:53.104736 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:53 crc kubenswrapper[4935]: I1005 08:42:53.499615 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"5e4a39ab2bd42fbc2828d6cf3524285a042b99f3b26230cc9f73dd09b7781035"} Oct 05 08:42:53 crc kubenswrapper[4935]: I1005 08:42:53.549792 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:53 crc kubenswrapper[4935]: I1005 08:42:53.608118 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-njgdw"] Oct 05 08:42:55 crc kubenswrapper[4935]: I1005 08:42:55.514779 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-njgdw" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerName="registry-server" containerID="cri-o://bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4" gracePeriod=2 Oct 05 08:42:55 crc kubenswrapper[4935]: I1005 08:42:55.931494 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:55 crc kubenswrapper[4935]: I1005 08:42:55.993749 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-catalog-content\") pod \"44d3b7ce-403d-437b-9c48-d44913198dc7\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " Oct 05 08:42:55 crc kubenswrapper[4935]: I1005 08:42:55.993829 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-utilities\") pod \"44d3b7ce-403d-437b-9c48-d44913198dc7\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " Oct 05 08:42:55 crc kubenswrapper[4935]: I1005 08:42:55.994060 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2pf8\" (UniqueName: \"kubernetes.io/projected/44d3b7ce-403d-437b-9c48-d44913198dc7-kube-api-access-k2pf8\") pod \"44d3b7ce-403d-437b-9c48-d44913198dc7\" (UID: \"44d3b7ce-403d-437b-9c48-d44913198dc7\") " Oct 05 08:42:55 crc kubenswrapper[4935]: I1005 08:42:55.995601 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-utilities" (OuterVolumeSpecName: "utilities") pod "44d3b7ce-403d-437b-9c48-d44913198dc7" (UID: "44d3b7ce-403d-437b-9c48-d44913198dc7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.006468 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44d3b7ce-403d-437b-9c48-d44913198dc7-kube-api-access-k2pf8" (OuterVolumeSpecName: "kube-api-access-k2pf8") pod "44d3b7ce-403d-437b-9c48-d44913198dc7" (UID: "44d3b7ce-403d-437b-9c48-d44913198dc7"). InnerVolumeSpecName "kube-api-access-k2pf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.097215 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2pf8\" (UniqueName: \"kubernetes.io/projected/44d3b7ce-403d-437b-9c48-d44913198dc7-kube-api-access-k2pf8\") on node \"crc\" DevicePath \"\"" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.097263 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.135757 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44d3b7ce-403d-437b-9c48-d44913198dc7" (UID: "44d3b7ce-403d-437b-9c48-d44913198dc7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.198003 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d3b7ce-403d-437b-9c48-d44913198dc7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.528782 4935 generic.go:334] "Generic (PLEG): container finished" podID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerID="bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4" exitCode=0 Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.528865 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-njgdw" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.528957 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njgdw" event={"ID":"44d3b7ce-403d-437b-9c48-d44913198dc7","Type":"ContainerDied","Data":"bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4"} Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.529567 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njgdw" event={"ID":"44d3b7ce-403d-437b-9c48-d44913198dc7","Type":"ContainerDied","Data":"2843cf977f2cbb6cdd23b947e8bb2fb66ff5484329dba528d497a0833f9fcb69"} Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.529616 4935 scope.go:117] "RemoveContainer" containerID="bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.553754 4935 scope.go:117] "RemoveContainer" containerID="8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.580814 4935 scope.go:117] "RemoveContainer" containerID="a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.589861 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-njgdw"] Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.603807 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-njgdw"] Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.629281 4935 scope.go:117] "RemoveContainer" containerID="bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4" Oct 05 08:42:56 crc kubenswrapper[4935]: E1005 08:42:56.629761 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4\": container with ID starting with bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4 not found: ID does not exist" containerID="bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.629804 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4"} err="failed to get container status \"bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4\": rpc error: code = NotFound desc = could not find container \"bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4\": container with ID starting with bda1096edcd6db04e5be0b808b5c4dc49701c392bb398d831508ce3ccc12a9d4 not found: ID does not exist" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.629830 4935 scope.go:117] "RemoveContainer" containerID="8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4" Oct 05 08:42:56 crc kubenswrapper[4935]: E1005 08:42:56.630261 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4\": container with ID starting with 8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4 not found: ID does not exist" containerID="8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.630284 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4"} err="failed to get container status \"8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4\": rpc error: code = NotFound desc = could not find container \"8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4\": container with ID starting with 8d64b44dcb13173c3f5b6814ecc5278adc25ec147e66ff6082b06f85cdc05be4 not found: ID does not exist" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.630302 4935 scope.go:117] "RemoveContainer" containerID="a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752" Oct 05 08:42:56 crc kubenswrapper[4935]: E1005 08:42:56.630736 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752\": container with ID starting with a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752 not found: ID does not exist" containerID="a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.630757 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752"} err="failed to get container status \"a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752\": rpc error: code = NotFound desc = could not find container \"a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752\": container with ID starting with a253128720e861a95dd9e7029c998008e42a27e2e78c266adc19801a6a73f752 not found: ID does not exist" Oct 05 08:42:56 crc kubenswrapper[4935]: I1005 08:42:56.788430 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" path="/var/lib/kubelet/pods/44d3b7ce-403d-437b-9c48-d44913198dc7/volumes" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.612921 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 08:43:31 crc kubenswrapper[4935]: E1005 08:43:31.613877 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerName="extract-utilities" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.614022 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerName="extract-utilities" Oct 05 08:43:31 crc kubenswrapper[4935]: E1005 08:43:31.614047 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerName="extract-content" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.614064 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerName="extract-content" Oct 05 08:43:31 crc kubenswrapper[4935]: E1005 08:43:31.614089 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerName="registry-server" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.614097 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerName="registry-server" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.614314 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="44d3b7ce-403d-437b-9c48-d44913198dc7" containerName="registry-server" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.615002 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.618500 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sw94x" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.635155 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.813800 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46f2r\" (UniqueName: \"kubernetes.io/projected/ebe6c9a7-c4a8-454a-973d-b6a1d849fdda-kube-api-access-46f2r\") pod \"mariadb-copy-data\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") " pod="openstack/mariadb-copy-data" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.814308 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\") pod \"mariadb-copy-data\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") " pod="openstack/mariadb-copy-data" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.916878 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46f2r\" (UniqueName: \"kubernetes.io/projected/ebe6c9a7-c4a8-454a-973d-b6a1d849fdda-kube-api-access-46f2r\") pod \"mariadb-copy-data\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") " pod="openstack/mariadb-copy-data" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.917336 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\") pod \"mariadb-copy-data\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") " pod="openstack/mariadb-copy-data" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.923544 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.923619 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\") pod \"mariadb-copy-data\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/77b1078059f6c2e2927d10f144208f839b04b6571a499e07cbbddf214038052d/globalmount\"" pod="openstack/mariadb-copy-data" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.957297 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46f2r\" (UniqueName: \"kubernetes.io/projected/ebe6c9a7-c4a8-454a-973d-b6a1d849fdda-kube-api-access-46f2r\") pod \"mariadb-copy-data\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") " pod="openstack/mariadb-copy-data" Oct 05 08:43:31 crc kubenswrapper[4935]: I1005 08:43:31.978106 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\") pod \"mariadb-copy-data\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") " pod="openstack/mariadb-copy-data" Oct 05 08:43:32 crc kubenswrapper[4935]: I1005 08:43:32.262138 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 05 08:43:32 crc kubenswrapper[4935]: I1005 08:43:32.640755 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 08:43:32 crc kubenswrapper[4935]: I1005 08:43:32.893110 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda","Type":"ContainerStarted","Data":"9818a1fd3c4ffb64ca1593726de0175068edda0a60422d2666f1be36fc7a773a"} Oct 05 08:43:32 crc kubenswrapper[4935]: I1005 08:43:32.893194 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda","Type":"ContainerStarted","Data":"7193eaff12ec2721972f59bddcc66171bca0b7c2f10275846f7579c66aeddc57"} Oct 05 08:43:35 crc kubenswrapper[4935]: I1005 08:43:35.760919 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=5.760883898 podStartE2EDuration="5.760883898s" podCreationTimestamp="2025-10-05 08:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:43:32.909443406 +0000 UTC m=+6646.792069866" watchObservedRunningTime="2025-10-05 08:43:35.760883898 +0000 UTC m=+6649.643510368" Oct 05 08:43:35 crc kubenswrapper[4935]: I1005 08:43:35.766701 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:35 crc kubenswrapper[4935]: I1005 08:43:35.767694 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:43:35 crc kubenswrapper[4935]: I1005 08:43:35.779501 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:35 crc kubenswrapper[4935]: I1005 08:43:35.889312 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrwdb\" (UniqueName: \"kubernetes.io/projected/9615a3e7-43c1-4acc-8d2a-9701ef10ead8-kube-api-access-vrwdb\") pod \"mariadb-client\" (UID: \"9615a3e7-43c1-4acc-8d2a-9701ef10ead8\") " pod="openstack/mariadb-client" Oct 05 08:43:35 crc kubenswrapper[4935]: I1005 08:43:35.991028 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrwdb\" (UniqueName: \"kubernetes.io/projected/9615a3e7-43c1-4acc-8d2a-9701ef10ead8-kube-api-access-vrwdb\") pod \"mariadb-client\" (UID: \"9615a3e7-43c1-4acc-8d2a-9701ef10ead8\") " pod="openstack/mariadb-client" Oct 05 08:43:36 crc kubenswrapper[4935]: I1005 08:43:36.023572 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrwdb\" (UniqueName: \"kubernetes.io/projected/9615a3e7-43c1-4acc-8d2a-9701ef10ead8-kube-api-access-vrwdb\") pod \"mariadb-client\" (UID: \"9615a3e7-43c1-4acc-8d2a-9701ef10ead8\") " pod="openstack/mariadb-client" Oct 05 08:43:36 crc kubenswrapper[4935]: I1005 08:43:36.098676 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:43:36 crc kubenswrapper[4935]: I1005 08:43:36.311591 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:36 crc kubenswrapper[4935]: I1005 08:43:36.937675 4935 generic.go:334] "Generic (PLEG): container finished" podID="9615a3e7-43c1-4acc-8d2a-9701ef10ead8" containerID="f07e806eb789958aaa2c066692ecdb7f619b9e11ecf096ba97aaab68aed4d23b" exitCode=0 Oct 05 08:43:36 crc kubenswrapper[4935]: I1005 08:43:36.937866 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"9615a3e7-43c1-4acc-8d2a-9701ef10ead8","Type":"ContainerDied","Data":"f07e806eb789958aaa2c066692ecdb7f619b9e11ecf096ba97aaab68aed4d23b"} Oct 05 08:43:36 crc kubenswrapper[4935]: I1005 08:43:36.937980 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"9615a3e7-43c1-4acc-8d2a-9701ef10ead8","Type":"ContainerStarted","Data":"1df0d5473127b0b23c5070a7dc21749d3b62742847aa1dd5b994846f2c3ac665"} Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.342820 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.370635 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_9615a3e7-43c1-4acc-8d2a-9701ef10ead8/mariadb-client/0.log" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.394812 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.404169 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.430517 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrwdb\" (UniqueName: \"kubernetes.io/projected/9615a3e7-43c1-4acc-8d2a-9701ef10ead8-kube-api-access-vrwdb\") pod \"9615a3e7-43c1-4acc-8d2a-9701ef10ead8\" (UID: \"9615a3e7-43c1-4acc-8d2a-9701ef10ead8\") " Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.436187 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9615a3e7-43c1-4acc-8d2a-9701ef10ead8-kube-api-access-vrwdb" (OuterVolumeSpecName: "kube-api-access-vrwdb") pod "9615a3e7-43c1-4acc-8d2a-9701ef10ead8" (UID: "9615a3e7-43c1-4acc-8d2a-9701ef10ead8"). InnerVolumeSpecName "kube-api-access-vrwdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.524738 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:38 crc kubenswrapper[4935]: E1005 08:43:38.525484 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9615a3e7-43c1-4acc-8d2a-9701ef10ead8" containerName="mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.525526 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9615a3e7-43c1-4acc-8d2a-9701ef10ead8" containerName="mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.525854 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9615a3e7-43c1-4acc-8d2a-9701ef10ead8" containerName="mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.526738 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.533063 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrwdb\" (UniqueName: \"kubernetes.io/projected/9615a3e7-43c1-4acc-8d2a-9701ef10ead8-kube-api-access-vrwdb\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.537023 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.635047 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkbbn\" (UniqueName: \"kubernetes.io/projected/2896728c-9111-4751-ba73-9ffe11d79e20-kube-api-access-hkbbn\") pod \"mariadb-client\" (UID: \"2896728c-9111-4751-ba73-9ffe11d79e20\") " pod="openstack/mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.736588 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkbbn\" (UniqueName: \"kubernetes.io/projected/2896728c-9111-4751-ba73-9ffe11d79e20-kube-api-access-hkbbn\") pod \"mariadb-client\" (UID: \"2896728c-9111-4751-ba73-9ffe11d79e20\") " pod="openstack/mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.757869 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkbbn\" (UniqueName: \"kubernetes.io/projected/2896728c-9111-4751-ba73-9ffe11d79e20-kube-api-access-hkbbn\") pod \"mariadb-client\" (UID: \"2896728c-9111-4751-ba73-9ffe11d79e20\") " pod="openstack/mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.812977 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9615a3e7-43c1-4acc-8d2a-9701ef10ead8" path="/var/lib/kubelet/pods/9615a3e7-43c1-4acc-8d2a-9701ef10ead8/volumes" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.853944 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.968620 4935 scope.go:117] "RemoveContainer" containerID="f07e806eb789958aaa2c066692ecdb7f619b9e11ecf096ba97aaab68aed4d23b" Oct 05 08:43:38 crc kubenswrapper[4935]: I1005 08:43:38.968649 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:43:39 crc kubenswrapper[4935]: I1005 08:43:39.349211 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:39 crc kubenswrapper[4935]: W1005 08:43:39.354572 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2896728c_9111_4751_ba73_9ffe11d79e20.slice/crio-fd40804bd476b82f99cd05ef02605f8c97c7a4bc8c5f02d1bc6207a8c9bcc27f WatchSource:0}: Error finding container fd40804bd476b82f99cd05ef02605f8c97c7a4bc8c5f02d1bc6207a8c9bcc27f: Status 404 returned error can't find the container with id fd40804bd476b82f99cd05ef02605f8c97c7a4bc8c5f02d1bc6207a8c9bcc27f Oct 05 08:43:39 crc kubenswrapper[4935]: I1005 08:43:39.980566 4935 generic.go:334] "Generic (PLEG): container finished" podID="2896728c-9111-4751-ba73-9ffe11d79e20" containerID="80a35954bc1fe1b70a90a2ca5f6ae6b64172125e8a9480bdf87409babde50cb1" exitCode=0 Oct 05 08:43:39 crc kubenswrapper[4935]: I1005 08:43:39.981022 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"2896728c-9111-4751-ba73-9ffe11d79e20","Type":"ContainerDied","Data":"80a35954bc1fe1b70a90a2ca5f6ae6b64172125e8a9480bdf87409babde50cb1"} Oct 05 08:43:39 crc kubenswrapper[4935]: I1005 08:43:39.981074 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"2896728c-9111-4751-ba73-9ffe11d79e20","Type":"ContainerStarted","Data":"fd40804bd476b82f99cd05ef02605f8c97c7a4bc8c5f02d1bc6207a8c9bcc27f"} Oct 05 08:43:41 crc kubenswrapper[4935]: I1005 08:43:41.378212 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:43:41 crc kubenswrapper[4935]: I1005 08:43:41.398742 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_2896728c-9111-4751-ba73-9ffe11d79e20/mariadb-client/0.log" Oct 05 08:43:41 crc kubenswrapper[4935]: I1005 08:43:41.434863 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:41 crc kubenswrapper[4935]: I1005 08:43:41.446045 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:43:41 crc kubenswrapper[4935]: I1005 08:43:41.480781 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkbbn\" (UniqueName: \"kubernetes.io/projected/2896728c-9111-4751-ba73-9ffe11d79e20-kube-api-access-hkbbn\") pod \"2896728c-9111-4751-ba73-9ffe11d79e20\" (UID: \"2896728c-9111-4751-ba73-9ffe11d79e20\") " Oct 05 08:43:41 crc kubenswrapper[4935]: I1005 08:43:41.489871 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2896728c-9111-4751-ba73-9ffe11d79e20-kube-api-access-hkbbn" (OuterVolumeSpecName: "kube-api-access-hkbbn") pod "2896728c-9111-4751-ba73-9ffe11d79e20" (UID: "2896728c-9111-4751-ba73-9ffe11d79e20"). InnerVolumeSpecName "kube-api-access-hkbbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:43:41 crc kubenswrapper[4935]: I1005 08:43:41.583120 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkbbn\" (UniqueName: \"kubernetes.io/projected/2896728c-9111-4751-ba73-9ffe11d79e20-kube-api-access-hkbbn\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:42 crc kubenswrapper[4935]: I1005 08:43:42.007218 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd40804bd476b82f99cd05ef02605f8c97c7a4bc8c5f02d1bc6207a8c9bcc27f" Oct 05 08:43:42 crc kubenswrapper[4935]: I1005 08:43:42.007349 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:43:42 crc kubenswrapper[4935]: I1005 08:43:42.795093 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2896728c-9111-4751-ba73-9ffe11d79e20" path="/var/lib/kubelet/pods/2896728c-9111-4751-ba73-9ffe11d79e20/volumes" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.115168 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8849h"] Oct 05 08:43:44 crc kubenswrapper[4935]: E1005 08:43:44.121680 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2896728c-9111-4751-ba73-9ffe11d79e20" containerName="mariadb-client" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.121724 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2896728c-9111-4751-ba73-9ffe11d79e20" containerName="mariadb-client" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.122085 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2896728c-9111-4751-ba73-9ffe11d79e20" containerName="mariadb-client" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.124723 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.131181 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8849h"] Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.231708 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-utilities\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.231769 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-kube-api-access-g5bxq\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.231796 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-catalog-content\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.333429 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-utilities\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.333488 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-kube-api-access-g5bxq\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.333533 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-catalog-content\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.334040 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-catalog-content\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.334050 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-utilities\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.353985 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-kube-api-access-g5bxq\") pod \"certified-operators-8849h\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.484350 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:44 crc kubenswrapper[4935]: I1005 08:43:44.962100 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8849h"] Oct 05 08:43:44 crc kubenswrapper[4935]: W1005 08:43:44.970117 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ccf2cd4_a7f7_405a_8eb5_d1240c1627e5.slice/crio-a7036f32c6f8691aefc9f2f9357d823a930ef7801f37e9adc0e6895d733ceae2 WatchSource:0}: Error finding container a7036f32c6f8691aefc9f2f9357d823a930ef7801f37e9adc0e6895d733ceae2: Status 404 returned error can't find the container with id a7036f32c6f8691aefc9f2f9357d823a930ef7801f37e9adc0e6895d733ceae2 Oct 05 08:43:45 crc kubenswrapper[4935]: I1005 08:43:45.035424 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8849h" event={"ID":"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5","Type":"ContainerStarted","Data":"a7036f32c6f8691aefc9f2f9357d823a930ef7801f37e9adc0e6895d733ceae2"} Oct 05 08:43:46 crc kubenswrapper[4935]: I1005 08:43:46.045982 4935 generic.go:334] "Generic (PLEG): container finished" podID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerID="be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d" exitCode=0 Oct 05 08:43:46 crc kubenswrapper[4935]: I1005 08:43:46.046107 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8849h" event={"ID":"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5","Type":"ContainerDied","Data":"be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d"} Oct 05 08:43:47 crc kubenswrapper[4935]: I1005 08:43:47.092703 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8849h" event={"ID":"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5","Type":"ContainerStarted","Data":"908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2"} Oct 05 08:43:48 crc kubenswrapper[4935]: I1005 08:43:48.104840 4935 generic.go:334] "Generic (PLEG): container finished" podID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerID="908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2" exitCode=0 Oct 05 08:43:48 crc kubenswrapper[4935]: I1005 08:43:48.104911 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8849h" event={"ID":"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5","Type":"ContainerDied","Data":"908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2"} Oct 05 08:43:49 crc kubenswrapper[4935]: I1005 08:43:49.115960 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8849h" event={"ID":"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5","Type":"ContainerStarted","Data":"1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b"} Oct 05 08:43:49 crc kubenswrapper[4935]: I1005 08:43:49.137339 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8849h" podStartSLOduration=2.646828456 podStartE2EDuration="5.137316948s" podCreationTimestamp="2025-10-05 08:43:44 +0000 UTC" firstStartedPulling="2025-10-05 08:43:46.047722929 +0000 UTC m=+6659.930349399" lastFinishedPulling="2025-10-05 08:43:48.538211411 +0000 UTC m=+6662.420837891" observedRunningTime="2025-10-05 08:43:49.135685155 +0000 UTC m=+6663.018311625" watchObservedRunningTime="2025-10-05 08:43:49.137316948 +0000 UTC m=+6663.019943408" Oct 05 08:43:54 crc kubenswrapper[4935]: I1005 08:43:54.484839 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:54 crc kubenswrapper[4935]: I1005 08:43:54.485476 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:54 crc kubenswrapper[4935]: I1005 08:43:54.539611 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:55 crc kubenswrapper[4935]: I1005 08:43:55.224282 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:55 crc kubenswrapper[4935]: I1005 08:43:55.282085 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8849h"] Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.183050 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8849h" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerName="registry-server" containerID="cri-o://1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b" gracePeriod=2 Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.569571 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.674862 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-kube-api-access-g5bxq\") pod \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.675304 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-catalog-content\") pod \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.675330 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-utilities\") pod \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\" (UID: \"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5\") " Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.676707 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-utilities" (OuterVolumeSpecName: "utilities") pod "9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" (UID: "9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.685099 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-kube-api-access-g5bxq" (OuterVolumeSpecName: "kube-api-access-g5bxq") pod "9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" (UID: "9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5"). InnerVolumeSpecName "kube-api-access-g5bxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.777198 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.777237 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-kube-api-access-g5bxq\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.778687 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" (UID: "9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:43:57 crc kubenswrapper[4935]: I1005 08:43:57.879468 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.192509 4935 generic.go:334] "Generic (PLEG): container finished" podID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerID="1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b" exitCode=0 Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.192573 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8849h" event={"ID":"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5","Type":"ContainerDied","Data":"1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b"} Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.192616 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8849h" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.192643 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8849h" event={"ID":"9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5","Type":"ContainerDied","Data":"a7036f32c6f8691aefc9f2f9357d823a930ef7801f37e9adc0e6895d733ceae2"} Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.192663 4935 scope.go:117] "RemoveContainer" containerID="1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.210455 4935 scope.go:117] "RemoveContainer" containerID="908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.236832 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8849h"] Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.239142 4935 scope.go:117] "RemoveContainer" containerID="be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.250553 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8849h"] Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.267770 4935 scope.go:117] "RemoveContainer" containerID="1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b" Oct 05 08:43:58 crc kubenswrapper[4935]: E1005 08:43:58.268378 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b\": container with ID starting with 1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b not found: ID does not exist" containerID="1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.268430 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b"} err="failed to get container status \"1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b\": rpc error: code = NotFound desc = could not find container \"1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b\": container with ID starting with 1bc5ead0ef67ac71dcfa787d6ecfbc53adb2038baf094731a93b8d8ff492a24b not found: ID does not exist" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.268453 4935 scope.go:117] "RemoveContainer" containerID="908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2" Oct 05 08:43:58 crc kubenswrapper[4935]: E1005 08:43:58.268886 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2\": container with ID starting with 908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2 not found: ID does not exist" containerID="908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.268962 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2"} err="failed to get container status \"908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2\": rpc error: code = NotFound desc = could not find container \"908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2\": container with ID starting with 908e27e214484abf2dbe692cc5ff3a4c1f1292edbef4bc31984e0408e00b9bf2 not found: ID does not exist" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.269006 4935 scope.go:117] "RemoveContainer" containerID="be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d" Oct 05 08:43:58 crc kubenswrapper[4935]: E1005 08:43:58.269364 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d\": container with ID starting with be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d not found: ID does not exist" containerID="be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.269397 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d"} err="failed to get container status \"be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d\": rpc error: code = NotFound desc = could not find container \"be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d\": container with ID starting with be1e8be56a4bbdc787de156f023a00e2cfb5376c4dbc6a8e5ee61e1f121d173d not found: ID does not exist" Oct 05 08:43:58 crc kubenswrapper[4935]: I1005 08:43:58.791130 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" path="/var/lib/kubelet/pods/9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5/volumes" Oct 05 08:44:08 crc kubenswrapper[4935]: E1005 08:44:08.109384 4935 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.59:60620->38.102.83.59:37063: write tcp 38.102.83.59:60620->38.102.83.59:37063: write: connection reset by peer Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.895700 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 08:44:10 crc kubenswrapper[4935]: E1005 08:44:10.896801 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerName="registry-server" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.896820 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerName="registry-server" Oct 05 08:44:10 crc kubenswrapper[4935]: E1005 08:44:10.896838 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerName="extract-utilities" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.896848 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerName="extract-utilities" Oct 05 08:44:10 crc kubenswrapper[4935]: E1005 08:44:10.896863 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerName="extract-content" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.896872 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerName="extract-content" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.897071 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ccf2cd4-a7f7-405a-8eb5-d1240c1627e5" containerName="registry-server" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.898004 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.900781 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.901027 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rrfhd" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.901718 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.910310 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.911594 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.923299 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.951626 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.964999 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.965140 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.990112 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.994557 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63ebf5e7-d07d-40af-890c-ecb794b3749c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.994626 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44p6l\" (UniqueName: \"kubernetes.io/projected/a258f8aa-2ec9-4098-b98f-23051ca75c08-kube-api-access-44p6l\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.994656 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqwpc\" (UniqueName: \"kubernetes.io/projected/63ebf5e7-d07d-40af-890c-ecb794b3749c-kube-api-access-dqwpc\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.994760 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63ebf5e7-d07d-40af-890c-ecb794b3749c-config\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.994829 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a258f8aa-2ec9-4098-b98f-23051ca75c08-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.994885 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ebf5e7-d07d-40af-890c-ecb794b3749c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.994956 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a258f8aa-2ec9-4098-b98f-23051ca75c08-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.995063 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a258f8aa-2ec9-4098-b98f-23051ca75c08-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.995099 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a258f8aa-2ec9-4098-b98f-23051ca75c08-config\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.995172 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-18b356dd-8a9d-47e3-a39f-73f6f9014cd3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b356dd-8a9d-47e3-a39f-73f6f9014cd3\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.995200 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8099b8d6-3f66-497f-a1eb-a990d1f10ea8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8099b8d6-3f66-497f-a1eb-a990d1f10ea8\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:10 crc kubenswrapper[4935]: I1005 08:44:10.995307 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63ebf5e7-d07d-40af-890c-ecb794b3749c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.096580 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmkrh\" (UniqueName: \"kubernetes.io/projected/3293245e-b8f2-4081-a526-227a9918dbe0-kube-api-access-nmkrh\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.096661 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44p6l\" (UniqueName: \"kubernetes.io/projected/a258f8aa-2ec9-4098-b98f-23051ca75c08-kube-api-access-44p6l\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.096698 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqwpc\" (UniqueName: \"kubernetes.io/projected/63ebf5e7-d07d-40af-890c-ecb794b3749c-kube-api-access-dqwpc\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.096751 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63ebf5e7-d07d-40af-890c-ecb794b3749c-config\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.096786 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a258f8aa-2ec9-4098-b98f-23051ca75c08-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.096825 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ebf5e7-d07d-40af-890c-ecb794b3749c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.096873 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a258f8aa-2ec9-4098-b98f-23051ca75c08-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.096965 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a258f8aa-2ec9-4098-b98f-23051ca75c08-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097008 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a258f8aa-2ec9-4098-b98f-23051ca75c08-config\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097062 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-18b356dd-8a9d-47e3-a39f-73f6f9014cd3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b356dd-8a9d-47e3-a39f-73f6f9014cd3\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097096 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8099b8d6-3f66-497f-a1eb-a990d1f10ea8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8099b8d6-3f66-497f-a1eb-a990d1f10ea8\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097164 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3293245e-b8f2-4081-a526-227a9918dbe0-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097191 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3293245e-b8f2-4081-a526-227a9918dbe0-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097218 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63ebf5e7-d07d-40af-890c-ecb794b3749c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097261 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3293245e-b8f2-4081-a526-227a9918dbe0-config\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097287 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3293245e-b8f2-4081-a526-227a9918dbe0-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097328 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63ebf5e7-d07d-40af-890c-ecb794b3749c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.097360 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-29dd3989-dcb5-4515-bd65-20d478d642d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29dd3989-dcb5-4515-bd65-20d478d642d6\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.099373 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63ebf5e7-d07d-40af-890c-ecb794b3749c-config\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.102982 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a258f8aa-2ec9-4098-b98f-23051ca75c08-config\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.103351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a258f8aa-2ec9-4098-b98f-23051ca75c08-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.103545 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a258f8aa-2ec9-4098-b98f-23051ca75c08-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.105217 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.108384 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63ebf5e7-d07d-40af-890c-ecb794b3749c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.108791 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63ebf5e7-d07d-40af-890c-ecb794b3749c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.109287 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.109329 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ebf5e7-d07d-40af-890c-ecb794b3749c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.112651 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.112678 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-18b356dd-8a9d-47e3-a39f-73f6f9014cd3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b356dd-8a9d-47e3-a39f-73f6f9014cd3\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f1a13cc34574afaf9d5877fb4f6ebba13667082c2853b3ff7195f2691ac1ccf1/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.112797 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.113243 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-rq8kt" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.113445 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.125976 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a258f8aa-2ec9-4098-b98f-23051ca75c08-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.127217 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqwpc\" (UniqueName: \"kubernetes.io/projected/63ebf5e7-d07d-40af-890c-ecb794b3749c-kube-api-access-dqwpc\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.134614 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44p6l\" (UniqueName: \"kubernetes.io/projected/a258f8aa-2ec9-4098-b98f-23051ca75c08-kube-api-access-44p6l\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.150246 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.153921 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.153970 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8099b8d6-3f66-497f-a1eb-a990d1f10ea8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8099b8d6-3f66-497f-a1eb-a990d1f10ea8\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9578847f06f54d86dceb85c96fa83f0d0a02657debea6808687046d52f290e42/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.170071 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.174509 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.185477 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.186881 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.191420 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8099b8d6-3f66-497f-a1eb-a990d1f10ea8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8099b8d6-3f66-497f-a1eb-a990d1f10ea8\") pod \"ovsdbserver-nb-0\" (UID: \"63ebf5e7-d07d-40af-890c-ecb794b3749c\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.198280 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.198712 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-aac8e5ea-cf6d-4697-8fae-460991dfb6a4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aac8e5ea-cf6d-4697-8fae-460991dfb6a4\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.198778 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419c6823-dbe7-4671-a7ea-ea7c91dff77d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.198816 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419c6823-dbe7-4671-a7ea-ea7c91dff77d-config\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.198920 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxqbk\" (UniqueName: \"kubernetes.io/projected/419c6823-dbe7-4671-a7ea-ea7c91dff77d-kube-api-access-nxqbk\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.199642 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/419c6823-dbe7-4671-a7ea-ea7c91dff77d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.199684 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3293245e-b8f2-4081-a526-227a9918dbe0-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.199704 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3293245e-b8f2-4081-a526-227a9918dbe0-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.199738 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3293245e-b8f2-4081-a526-227a9918dbe0-config\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.199759 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3293245e-b8f2-4081-a526-227a9918dbe0-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.199787 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-29dd3989-dcb5-4515-bd65-20d478d642d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29dd3989-dcb5-4515-bd65-20d478d642d6\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.199807 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/419c6823-dbe7-4671-a7ea-ea7c91dff77d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.199859 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmkrh\" (UniqueName: \"kubernetes.io/projected/3293245e-b8f2-4081-a526-227a9918dbe0-kube-api-access-nmkrh\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.200763 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3293245e-b8f2-4081-a526-227a9918dbe0-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.202791 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3293245e-b8f2-4081-a526-227a9918dbe0-config\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.203542 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3293245e-b8f2-4081-a526-227a9918dbe0-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.203572 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.203601 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-29dd3989-dcb5-4515-bd65-20d478d642d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29dd3989-dcb5-4515-bd65-20d478d642d6\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/674373bcc24c2d6d3425728f84f6eb448438cd255bce35d5288050f88f817e30/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.205216 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3293245e-b8f2-4081-a526-227a9918dbe0-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.206580 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-18b356dd-8a9d-47e3-a39f-73f6f9014cd3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b356dd-8a9d-47e3-a39f-73f6f9014cd3\") pod \"ovsdbserver-nb-2\" (UID: \"a258f8aa-2ec9-4098-b98f-23051ca75c08\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.209056 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.217147 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmkrh\" (UniqueName: \"kubernetes.io/projected/3293245e-b8f2-4081-a526-227a9918dbe0-kube-api-access-nmkrh\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.223010 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.238305 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.242004 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-29dd3989-dcb5-4515-bd65-20d478d642d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29dd3989-dcb5-4515-bd65-20d478d642d6\") pod \"ovsdbserver-nb-1\" (UID: \"3293245e-b8f2-4081-a526-227a9918dbe0\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.285084 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301225 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4315a8d6-2d0a-4576-be77-30e0b2096f0f-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301535 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4315a8d6-2d0a-4576-be77-30e0b2096f0f-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301565 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94cd3f0f-4346-424d-9ea9-4d8153d7d418-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301590 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4315a8d6-2d0a-4576-be77-30e0b2096f0f-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301622 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8c474628-e8b5-4be3-ae42-d4c481f3df72\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c474628-e8b5-4be3-ae42-d4c481f3df72\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301650 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94cd3f0f-4346-424d-9ea9-4d8153d7d418-config\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301683 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419c6823-dbe7-4671-a7ea-ea7c91dff77d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301708 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-567km\" (UniqueName: \"kubernetes.io/projected/94cd3f0f-4346-424d-9ea9-4d8153d7d418-kube-api-access-567km\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301815 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419c6823-dbe7-4671-a7ea-ea7c91dff77d-config\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301924 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxqbk\" (UniqueName: \"kubernetes.io/projected/419c6823-dbe7-4671-a7ea-ea7c91dff77d-kube-api-access-nxqbk\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.301968 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/419c6823-dbe7-4671-a7ea-ea7c91dff77d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.302019 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfnbw\" (UniqueName: \"kubernetes.io/projected/4315a8d6-2d0a-4576-be77-30e0b2096f0f-kube-api-access-mfnbw\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.302117 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94cd3f0f-4346-424d-9ea9-4d8153d7d418-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.302150 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94cd3f0f-4346-424d-9ea9-4d8153d7d418-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.302171 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4315a8d6-2d0a-4576-be77-30e0b2096f0f-config\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.302241 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/419c6823-dbe7-4671-a7ea-ea7c91dff77d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.302269 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-47277c0b-a015-417a-9374-2c460f056a0f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47277c0b-a015-417a-9374-2c460f056a0f\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.302355 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-aac8e5ea-cf6d-4697-8fae-460991dfb6a4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aac8e5ea-cf6d-4697-8fae-460991dfb6a4\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.303094 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419c6823-dbe7-4671-a7ea-ea7c91dff77d-config\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.303608 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/419c6823-dbe7-4671-a7ea-ea7c91dff77d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.305462 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/419c6823-dbe7-4671-a7ea-ea7c91dff77d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.305490 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419c6823-dbe7-4671-a7ea-ea7c91dff77d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.307852 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.307908 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-aac8e5ea-cf6d-4697-8fae-460991dfb6a4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aac8e5ea-cf6d-4697-8fae-460991dfb6a4\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/118ef0221af7e3ffa90947d2095d0ad96bd5bc37f50f6a76780a620dc29f8de2/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.326708 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxqbk\" (UniqueName: \"kubernetes.io/projected/419c6823-dbe7-4671-a7ea-ea7c91dff77d-kube-api-access-nxqbk\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.368656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-aac8e5ea-cf6d-4697-8fae-460991dfb6a4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aac8e5ea-cf6d-4697-8fae-460991dfb6a4\") pod \"ovsdbserver-sb-0\" (UID: \"419c6823-dbe7-4671-a7ea-ea7c91dff77d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406116 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94cd3f0f-4346-424d-9ea9-4d8153d7d418-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406167 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94cd3f0f-4346-424d-9ea9-4d8153d7d418-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406191 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4315a8d6-2d0a-4576-be77-30e0b2096f0f-config\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406236 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-47277c0b-a015-417a-9374-2c460f056a0f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47277c0b-a015-417a-9374-2c460f056a0f\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406364 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4315a8d6-2d0a-4576-be77-30e0b2096f0f-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406393 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4315a8d6-2d0a-4576-be77-30e0b2096f0f-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406417 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94cd3f0f-4346-424d-9ea9-4d8153d7d418-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406441 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4315a8d6-2d0a-4576-be77-30e0b2096f0f-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406463 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8c474628-e8b5-4be3-ae42-d4c481f3df72\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c474628-e8b5-4be3-ae42-d4c481f3df72\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406484 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94cd3f0f-4346-424d-9ea9-4d8153d7d418-config\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406529 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-567km\" (UniqueName: \"kubernetes.io/projected/94cd3f0f-4346-424d-9ea9-4d8153d7d418-kube-api-access-567km\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.406611 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfnbw\" (UniqueName: \"kubernetes.io/projected/4315a8d6-2d0a-4576-be77-30e0b2096f0f-kube-api-access-mfnbw\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.407380 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4315a8d6-2d0a-4576-be77-30e0b2096f0f-config\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.407815 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4315a8d6-2d0a-4576-be77-30e0b2096f0f-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.407843 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94cd3f0f-4346-424d-9ea9-4d8153d7d418-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.408084 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94cd3f0f-4346-424d-9ea9-4d8153d7d418-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.408545 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94cd3f0f-4346-424d-9ea9-4d8153d7d418-config\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.408883 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4315a8d6-2d0a-4576-be77-30e0b2096f0f-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.413373 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.413413 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8c474628-e8b5-4be3-ae42-d4c481f3df72\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c474628-e8b5-4be3-ae42-d4c481f3df72\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b62484c846640226e008936c1e3382139b7e156c45b61fb8f18bee68a090d641/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.413533 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.413574 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-47277c0b-a015-417a-9374-2c460f056a0f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47277c0b-a015-417a-9374-2c460f056a0f\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/94cc7e37060fcb9d4be5a3e9e7ad10281c509eac6ac873aaa4a83e5b19f146ad/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.415003 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94cd3f0f-4346-424d-9ea9-4d8153d7d418-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.415537 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4315a8d6-2d0a-4576-be77-30e0b2096f0f-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.427598 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfnbw\" (UniqueName: \"kubernetes.io/projected/4315a8d6-2d0a-4576-be77-30e0b2096f0f-kube-api-access-mfnbw\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.428191 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-567km\" (UniqueName: \"kubernetes.io/projected/94cd3f0f-4346-424d-9ea9-4d8153d7d418-kube-api-access-567km\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.443502 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8c474628-e8b5-4be3-ae42-d4c481f3df72\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c474628-e8b5-4be3-ae42-d4c481f3df72\") pod \"ovsdbserver-sb-1\" (UID: \"4315a8d6-2d0a-4576-be77-30e0b2096f0f\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.454568 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-47277c0b-a015-417a-9374-2c460f056a0f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47277c0b-a015-417a-9374-2c460f056a0f\") pod \"ovsdbserver-sb-2\" (UID: \"94cd3f0f-4346-424d-9ea9-4d8153d7d418\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.655718 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.674770 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.688600 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.796033 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.885128 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 05 08:44:11 crc kubenswrapper[4935]: W1005 08:44:11.898346 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda258f8aa_2ec9_4098_b98f_23051ca75c08.slice/crio-87b6c517fbfa2f1097cae1a99ee5e1f5690478bbbd6c9d797ce3ca9ed27a398f WatchSource:0}: Error finding container 87b6c517fbfa2f1097cae1a99ee5e1f5690478bbbd6c9d797ce3ca9ed27a398f: Status 404 returned error can't find the container with id 87b6c517fbfa2f1097cae1a99ee5e1f5690478bbbd6c9d797ce3ca9ed27a398f Oct 05 08:44:11 crc kubenswrapper[4935]: I1005 08:44:11.986875 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 08:44:11 crc kubenswrapper[4935]: W1005 08:44:11.990687 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod419c6823_dbe7_4671_a7ea_ea7c91dff77d.slice/crio-e92ed857a3ea712c14a128dfce17b20cae1a63356b28931700e0edc7bd4a0244 WatchSource:0}: Error finding container e92ed857a3ea712c14a128dfce17b20cae1a63356b28931700e0edc7bd4a0244: Status 404 returned error can't find the container with id e92ed857a3ea712c14a128dfce17b20cae1a63356b28931700e0edc7bd4a0244 Oct 05 08:44:12 crc kubenswrapper[4935]: I1005 08:44:12.341845 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 05 08:44:12 crc kubenswrapper[4935]: I1005 08:44:12.360850 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"94cd3f0f-4346-424d-9ea9-4d8153d7d418","Type":"ContainerStarted","Data":"de26506f9ffaf48c4ca34a2fa02ae3f92696995fc8560d03032e79cc1700be49"} Oct 05 08:44:12 crc kubenswrapper[4935]: I1005 08:44:12.364627 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"63ebf5e7-d07d-40af-890c-ecb794b3749c","Type":"ContainerStarted","Data":"e6edb1a823b576daadfdaa1d6e220e34cc361d7514fbdd6466131d671f510278"} Oct 05 08:44:12 crc kubenswrapper[4935]: I1005 08:44:12.367100 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"419c6823-dbe7-4671-a7ea-ea7c91dff77d","Type":"ContainerStarted","Data":"e92ed857a3ea712c14a128dfce17b20cae1a63356b28931700e0edc7bd4a0244"} Oct 05 08:44:12 crc kubenswrapper[4935]: I1005 08:44:12.368607 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a258f8aa-2ec9-4098-b98f-23051ca75c08","Type":"ContainerStarted","Data":"87b6c517fbfa2f1097cae1a99ee5e1f5690478bbbd6c9d797ce3ca9ed27a398f"} Oct 05 08:44:12 crc kubenswrapper[4935]: I1005 08:44:12.513815 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 05 08:44:12 crc kubenswrapper[4935]: W1005 08:44:12.519923 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3293245e_b8f2_4081_a526_227a9918dbe0.slice/crio-ee5b51a29380ac910c48b0385eac68b62b9c893b26b940178c5db1332bbd0d1f WatchSource:0}: Error finding container ee5b51a29380ac910c48b0385eac68b62b9c893b26b940178c5db1332bbd0d1f: Status 404 returned error can't find the container with id ee5b51a29380ac910c48b0385eac68b62b9c893b26b940178c5db1332bbd0d1f Oct 05 08:44:13 crc kubenswrapper[4935]: I1005 08:44:13.067632 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 05 08:44:13 crc kubenswrapper[4935]: W1005 08:44:13.072189 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4315a8d6_2d0a_4576_be77_30e0b2096f0f.slice/crio-84c8e72ec4de7b07a154c30541df65b6d264d9aca584086e97eaf89742aba60c WatchSource:0}: Error finding container 84c8e72ec4de7b07a154c30541df65b6d264d9aca584086e97eaf89742aba60c: Status 404 returned error can't find the container with id 84c8e72ec4de7b07a154c30541df65b6d264d9aca584086e97eaf89742aba60c Oct 05 08:44:13 crc kubenswrapper[4935]: I1005 08:44:13.377230 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"4315a8d6-2d0a-4576-be77-30e0b2096f0f","Type":"ContainerStarted","Data":"84c8e72ec4de7b07a154c30541df65b6d264d9aca584086e97eaf89742aba60c"} Oct 05 08:44:13 crc kubenswrapper[4935]: I1005 08:44:13.378719 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"3293245e-b8f2-4081-a526-227a9918dbe0","Type":"ContainerStarted","Data":"ee5b51a29380ac910c48b0385eac68b62b9c893b26b940178c5db1332bbd0d1f"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.409353 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"94cd3f0f-4346-424d-9ea9-4d8153d7d418","Type":"ContainerStarted","Data":"69290aa1a250e154b90955c29d64e15ca54b5c28d0899edfea5802af33f0cd81"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.410009 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"94cd3f0f-4346-424d-9ea9-4d8153d7d418","Type":"ContainerStarted","Data":"4703718e6e7aa8959f120720adaf4a17d71e75ca7f7005d0f834362c89cee61e"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.411914 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"3293245e-b8f2-4081-a526-227a9918dbe0","Type":"ContainerStarted","Data":"567fff8d0f38fa5f09c2b75be51eb4916dddf204e6a7fe51325eb1060767f09d"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.411945 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"3293245e-b8f2-4081-a526-227a9918dbe0","Type":"ContainerStarted","Data":"522db89a1f61ddf72013b208f1d87c82409f6178f24cb3ff978f8996548ac18e"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.414706 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"63ebf5e7-d07d-40af-890c-ecb794b3749c","Type":"ContainerStarted","Data":"b87a5c7de103a1f1a32074c33430a43c14c3a0989eba9ffc14a2b4e8db0dad1f"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.414761 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"63ebf5e7-d07d-40af-890c-ecb794b3749c","Type":"ContainerStarted","Data":"5cbf77aac3563264dea91b9c4d321f6edd8a3dd300721b4f40ae1a066aef471c"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.417391 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"419c6823-dbe7-4671-a7ea-ea7c91dff77d","Type":"ContainerStarted","Data":"08490ef38a605fca2851b590f7f9caf8f1d6335b0696d8e4d5271d229e9eee70"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.417432 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"419c6823-dbe7-4671-a7ea-ea7c91dff77d","Type":"ContainerStarted","Data":"f12ceadeb9548a3e8016b650b5823f5e53d480d5be15ca35974ee75a3fc2f42c"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.419804 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"4315a8d6-2d0a-4576-be77-30e0b2096f0f","Type":"ContainerStarted","Data":"7f0206ccc04d72be2545ea36afd89f8152ea692a227a1dd5795ac8a4fb00c710"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.419848 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"4315a8d6-2d0a-4576-be77-30e0b2096f0f","Type":"ContainerStarted","Data":"191dc65b38c4f8b290441187fb8bcbed155f1b58a30429fe2822dc423820ead3"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.423304 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a258f8aa-2ec9-4098-b98f-23051ca75c08","Type":"ContainerStarted","Data":"9a09b1599439c82744d30f59040601b75155ab63cdd2b93e0de5c6115647ba3a"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.423355 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a258f8aa-2ec9-4098-b98f-23051ca75c08","Type":"ContainerStarted","Data":"b88237d9b23432030bd6d076410a44f51e3e67c03d1bb5c68c7f44b2503aa057"} Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.436806 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.6616792719999998 podStartE2EDuration="7.436788358s" podCreationTimestamp="2025-10-05 08:44:10 +0000 UTC" firstStartedPulling="2025-10-05 08:44:12.332606393 +0000 UTC m=+6686.215232853" lastFinishedPulling="2025-10-05 08:44:16.107715469 +0000 UTC m=+6689.990341939" observedRunningTime="2025-10-05 08:44:17.429492363 +0000 UTC m=+6691.312118853" watchObservedRunningTime="2025-10-05 08:44:17.436788358 +0000 UTC m=+6691.319414828" Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.453927 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.178012156 podStartE2EDuration="8.453908683s" podCreationTimestamp="2025-10-05 08:44:09 +0000 UTC" firstStartedPulling="2025-10-05 08:44:11.83135592 +0000 UTC m=+6685.713982380" lastFinishedPulling="2025-10-05 08:44:16.107252447 +0000 UTC m=+6689.989878907" observedRunningTime="2025-10-05 08:44:17.449730242 +0000 UTC m=+6691.332356732" watchObservedRunningTime="2025-10-05 08:44:17.453908683 +0000 UTC m=+6691.336535153" Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.468642 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.356963951 podStartE2EDuration="7.468609295s" podCreationTimestamp="2025-10-05 08:44:10 +0000 UTC" firstStartedPulling="2025-10-05 08:44:11.998282824 +0000 UTC m=+6685.880909284" lastFinishedPulling="2025-10-05 08:44:16.109928118 +0000 UTC m=+6689.992554628" observedRunningTime="2025-10-05 08:44:17.467432663 +0000 UTC m=+6691.350059143" watchObservedRunningTime="2025-10-05 08:44:17.468609295 +0000 UTC m=+6691.351235795" Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.489538 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.431725638 podStartE2EDuration="7.489522161s" podCreationTimestamp="2025-10-05 08:44:10 +0000 UTC" firstStartedPulling="2025-10-05 08:44:13.074333536 +0000 UTC m=+6686.956959996" lastFinishedPulling="2025-10-05 08:44:16.132130059 +0000 UTC m=+6690.014756519" observedRunningTime="2025-10-05 08:44:17.489165632 +0000 UTC m=+6691.371792102" watchObservedRunningTime="2025-10-05 08:44:17.489522161 +0000 UTC m=+6691.372148621" Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.521010 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.877026523 podStartE2EDuration="8.520990459s" podCreationTimestamp="2025-10-05 08:44:09 +0000 UTC" firstStartedPulling="2025-10-05 08:44:12.522774864 +0000 UTC m=+6686.405401324" lastFinishedPulling="2025-10-05 08:44:16.1667388 +0000 UTC m=+6690.049365260" observedRunningTime="2025-10-05 08:44:17.515926434 +0000 UTC m=+6691.398552904" watchObservedRunningTime="2025-10-05 08:44:17.520990459 +0000 UTC m=+6691.403616919" Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.539549 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=4.337289506 podStartE2EDuration="8.539527972s" podCreationTimestamp="2025-10-05 08:44:09 +0000 UTC" firstStartedPulling="2025-10-05 08:44:11.90609193 +0000 UTC m=+6685.788718380" lastFinishedPulling="2025-10-05 08:44:16.108330386 +0000 UTC m=+6689.990956846" observedRunningTime="2025-10-05 08:44:17.534343584 +0000 UTC m=+6691.416970044" watchObservedRunningTime="2025-10-05 08:44:17.539527972 +0000 UTC m=+6691.422154432" Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.655832 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.675101 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:17 crc kubenswrapper[4935]: I1005 08:44:17.689386 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.223411 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.239926 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.266847 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.286212 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.327298 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.350921 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.456497 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.457007 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.457028 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.726663 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.728365 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.761351 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.761963 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.762138 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:20 crc kubenswrapper[4935]: I1005 08:44:20.762472 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.295123 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.323612 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.362938 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.522453 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.526590 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.527410 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.536583 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f55dcf55c-qjfj6"] Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.537992 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.554694 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.557491 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f55dcf55c-qjfj6"] Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.589874 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh9sf\" (UniqueName: \"kubernetes.io/projected/968afd10-dc17-47a0-b533-5e451e91e2c8-kube-api-access-hh9sf\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.589986 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-ovsdbserver-nb\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.590025 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-dns-svc\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.590049 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-config\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.691773 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-ovsdbserver-nb\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.691875 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-dns-svc\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.691937 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-config\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.692183 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh9sf\" (UniqueName: \"kubernetes.io/projected/968afd10-dc17-47a0-b533-5e451e91e2c8-kube-api-access-hh9sf\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.692939 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-ovsdbserver-nb\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.693265 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-dns-svc\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.693470 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-config\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.718041 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh9sf\" (UniqueName: \"kubernetes.io/projected/968afd10-dc17-47a0-b533-5e451e91e2c8-kube-api-access-hh9sf\") pod \"dnsmasq-dns-f55dcf55c-qjfj6\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.871407 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.912121 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f55dcf55c-qjfj6"] Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.953905 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b486dfb5c-xkdlz"] Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.958047 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:21 crc kubenswrapper[4935]: I1005 08:44:21.960967 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.018723 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b486dfb5c-xkdlz"] Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.097681 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-sb\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.098025 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddpqd\" (UniqueName: \"kubernetes.io/projected/957fa2eb-5dc0-40c2-9252-2161d11ce895-kube-api-access-ddpqd\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.098103 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-dns-svc\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.098141 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-config\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.098208 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-nb\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.200774 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-dns-svc\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.200847 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-config\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.200932 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-nb\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.200991 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-sb\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.201015 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddpqd\" (UniqueName: \"kubernetes.io/projected/957fa2eb-5dc0-40c2-9252-2161d11ce895-kube-api-access-ddpqd\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.202110 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-nb\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.202166 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-config\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.202177 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-sb\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.203139 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-dns-svc\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.222963 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddpqd\" (UniqueName: \"kubernetes.io/projected/957fa2eb-5dc0-40c2-9252-2161d11ce895-kube-api-access-ddpqd\") pod \"dnsmasq-dns-5b486dfb5c-xkdlz\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.326215 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.373099 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f55dcf55c-qjfj6"] Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.477394 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" event={"ID":"968afd10-dc17-47a0-b533-5e451e91e2c8","Type":"ContainerStarted","Data":"2ad66a35849a04b864ca64219f2ae21b9e6de5292410f4f6742d469bc28ffbce"} Oct 05 08:44:22 crc kubenswrapper[4935]: W1005 08:44:22.799487 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod957fa2eb_5dc0_40c2_9252_2161d11ce895.slice/crio-2bace3c8ad78200307dfa3f1baa907c6d7235e1c6483d46bb6dcc28b35d1192e WatchSource:0}: Error finding container 2bace3c8ad78200307dfa3f1baa907c6d7235e1c6483d46bb6dcc28b35d1192e: Status 404 returned error can't find the container with id 2bace3c8ad78200307dfa3f1baa907c6d7235e1c6483d46bb6dcc28b35d1192e Oct 05 08:44:22 crc kubenswrapper[4935]: I1005 08:44:22.800115 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b486dfb5c-xkdlz"] Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.494724 4935 generic.go:334] "Generic (PLEG): container finished" podID="968afd10-dc17-47a0-b533-5e451e91e2c8" containerID="01820e2e658e8cdd81280b0f1f6314fba9cb915b8035abc102c115288cdee0e2" exitCode=0 Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.494787 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" event={"ID":"968afd10-dc17-47a0-b533-5e451e91e2c8","Type":"ContainerDied","Data":"01820e2e658e8cdd81280b0f1f6314fba9cb915b8035abc102c115288cdee0e2"} Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.506434 4935 generic.go:334] "Generic (PLEG): container finished" podID="957fa2eb-5dc0-40c2-9252-2161d11ce895" containerID="8d4454891d58dc3205d7d1008a031f209f6a1af65a250b8466835c51e55a4115" exitCode=0 Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.506523 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" event={"ID":"957fa2eb-5dc0-40c2-9252-2161d11ce895","Type":"ContainerDied","Data":"8d4454891d58dc3205d7d1008a031f209f6a1af65a250b8466835c51e55a4115"} Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.506568 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" event={"ID":"957fa2eb-5dc0-40c2-9252-2161d11ce895","Type":"ContainerStarted","Data":"2bace3c8ad78200307dfa3f1baa907c6d7235e1c6483d46bb6dcc28b35d1192e"} Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.855188 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.932749 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-ovsdbserver-nb\") pod \"968afd10-dc17-47a0-b533-5e451e91e2c8\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.932821 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-config\") pod \"968afd10-dc17-47a0-b533-5e451e91e2c8\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.932945 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-dns-svc\") pod \"968afd10-dc17-47a0-b533-5e451e91e2c8\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.933030 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh9sf\" (UniqueName: \"kubernetes.io/projected/968afd10-dc17-47a0-b533-5e451e91e2c8-kube-api-access-hh9sf\") pod \"968afd10-dc17-47a0-b533-5e451e91e2c8\" (UID: \"968afd10-dc17-47a0-b533-5e451e91e2c8\") " Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.939634 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/968afd10-dc17-47a0-b533-5e451e91e2c8-kube-api-access-hh9sf" (OuterVolumeSpecName: "kube-api-access-hh9sf") pod "968afd10-dc17-47a0-b533-5e451e91e2c8" (UID: "968afd10-dc17-47a0-b533-5e451e91e2c8"). InnerVolumeSpecName "kube-api-access-hh9sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.951349 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-config" (OuterVolumeSpecName: "config") pod "968afd10-dc17-47a0-b533-5e451e91e2c8" (UID: "968afd10-dc17-47a0-b533-5e451e91e2c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.951822 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "968afd10-dc17-47a0-b533-5e451e91e2c8" (UID: "968afd10-dc17-47a0-b533-5e451e91e2c8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:44:23 crc kubenswrapper[4935]: I1005 08:44:23.952136 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "968afd10-dc17-47a0-b533-5e451e91e2c8" (UID: "968afd10-dc17-47a0-b533-5e451e91e2c8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.035926 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.035994 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.036014 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/968afd10-dc17-47a0-b533-5e451e91e2c8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.036032 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh9sf\" (UniqueName: \"kubernetes.io/projected/968afd10-dc17-47a0-b533-5e451e91e2c8-kube-api-access-hh9sf\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.522306 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" event={"ID":"968afd10-dc17-47a0-b533-5e451e91e2c8","Type":"ContainerDied","Data":"2ad66a35849a04b864ca64219f2ae21b9e6de5292410f4f6742d469bc28ffbce"} Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.522399 4935 scope.go:117] "RemoveContainer" containerID="01820e2e658e8cdd81280b0f1f6314fba9cb915b8035abc102c115288cdee0e2" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.522400 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f55dcf55c-qjfj6" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.527017 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" event={"ID":"957fa2eb-5dc0-40c2-9252-2161d11ce895","Type":"ContainerStarted","Data":"ae1dab29c65968860231c27e67156bdb9501a55e08cd293fa1e400be513d97f3"} Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.527179 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.528277 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 05 08:44:24 crc kubenswrapper[4935]: E1005 08:44:24.528656 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="968afd10-dc17-47a0-b533-5e451e91e2c8" containerName="init" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.528679 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="968afd10-dc17-47a0-b533-5e451e91e2c8" containerName="init" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.528879 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="968afd10-dc17-47a0-b533-5e451e91e2c8" containerName="init" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.529571 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.533130 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.547630 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.606179 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" podStartSLOduration=3.606152695 podStartE2EDuration="3.606152695s" podCreationTimestamp="2025-10-05 08:44:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:44:24.584410036 +0000 UTC m=+6698.467036536" watchObservedRunningTime="2025-10-05 08:44:24.606152695 +0000 UTC m=+6698.488779165" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.645263 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/ec474e79-9762-42c6-a931-bfa5896f7f00-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.645334 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.645451 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hx96\" (UniqueName: \"kubernetes.io/projected/ec474e79-9762-42c6-a931-bfa5896f7f00-kube-api-access-9hx96\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.649035 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f55dcf55c-qjfj6"] Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.657864 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f55dcf55c-qjfj6"] Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.747762 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hx96\" (UniqueName: \"kubernetes.io/projected/ec474e79-9762-42c6-a931-bfa5896f7f00-kube-api-access-9hx96\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.747928 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/ec474e79-9762-42c6-a931-bfa5896f7f00-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.747980 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.750006 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.750048 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1ba27b6bebbeefbedc10b5ad63ed89e2c696d6f38284b89516cd650d5d55c1df/globalmount\"" pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.751752 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/ec474e79-9762-42c6-a931-bfa5896f7f00-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.780557 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hx96\" (UniqueName: \"kubernetes.io/projected/ec474e79-9762-42c6-a931-bfa5896f7f00-kube-api-access-9hx96\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.788920 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="968afd10-dc17-47a0-b533-5e451e91e2c8" path="/var/lib/kubelet/pods/968afd10-dc17-47a0-b533-5e451e91e2c8/volumes" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.789489 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\") pod \"ovn-copy-data\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " pod="openstack/ovn-copy-data" Oct 05 08:44:24 crc kubenswrapper[4935]: I1005 08:44:24.890006 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 05 08:44:25 crc kubenswrapper[4935]: I1005 08:44:25.432145 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 08:44:25 crc kubenswrapper[4935]: I1005 08:44:25.539274 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"ec474e79-9762-42c6-a931-bfa5896f7f00","Type":"ContainerStarted","Data":"27f47c837e5a78c1d78398006072c5f31b7f66c33fe7c1605fc6cb22f70a2e88"} Oct 05 08:44:26 crc kubenswrapper[4935]: I1005 08:44:26.553554 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"ec474e79-9762-42c6-a931-bfa5896f7f00","Type":"ContainerStarted","Data":"17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66"} Oct 05 08:44:32 crc kubenswrapper[4935]: I1005 08:44:32.328807 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:44:32 crc kubenswrapper[4935]: I1005 08:44:32.366911 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=9.146010253 podStartE2EDuration="9.366855592s" podCreationTimestamp="2025-10-05 08:44:23 +0000 UTC" firstStartedPulling="2025-10-05 08:44:25.444673135 +0000 UTC m=+6699.327299595" lastFinishedPulling="2025-10-05 08:44:25.665518454 +0000 UTC m=+6699.548144934" observedRunningTime="2025-10-05 08:44:26.575548618 +0000 UTC m=+6700.458175118" watchObservedRunningTime="2025-10-05 08:44:32.366855592 +0000 UTC m=+6706.249482132" Oct 05 08:44:32 crc kubenswrapper[4935]: I1005 08:44:32.403334 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-n6qwp"] Oct 05 08:44:32 crc kubenswrapper[4935]: I1005 08:44:32.403788 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" podUID="60e2b0c1-83ce-4061-9baa-351e27b30b09" containerName="dnsmasq-dns" containerID="cri-o://2b4f1aff70ad057031684c6d6c77872d598fe45472224d32a3ef39d7a4bd3297" gracePeriod=10 Oct 05 08:44:32 crc kubenswrapper[4935]: I1005 08:44:32.616497 4935 generic.go:334] "Generic (PLEG): container finished" podID="60e2b0c1-83ce-4061-9baa-351e27b30b09" containerID="2b4f1aff70ad057031684c6d6c77872d598fe45472224d32a3ef39d7a4bd3297" exitCode=0 Oct 05 08:44:32 crc kubenswrapper[4935]: I1005 08:44:32.616552 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" event={"ID":"60e2b0c1-83ce-4061-9baa-351e27b30b09","Type":"ContainerDied","Data":"2b4f1aff70ad057031684c6d6c77872d598fe45472224d32a3ef39d7a4bd3297"} Oct 05 08:44:32 crc kubenswrapper[4935]: I1005 08:44:32.892096 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.003323 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j585\" (UniqueName: \"kubernetes.io/projected/60e2b0c1-83ce-4061-9baa-351e27b30b09-kube-api-access-8j585\") pod \"60e2b0c1-83ce-4061-9baa-351e27b30b09\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.003403 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-dns-svc\") pod \"60e2b0c1-83ce-4061-9baa-351e27b30b09\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.003427 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-config\") pod \"60e2b0c1-83ce-4061-9baa-351e27b30b09\" (UID: \"60e2b0c1-83ce-4061-9baa-351e27b30b09\") " Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.009986 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60e2b0c1-83ce-4061-9baa-351e27b30b09-kube-api-access-8j585" (OuterVolumeSpecName: "kube-api-access-8j585") pod "60e2b0c1-83ce-4061-9baa-351e27b30b09" (UID: "60e2b0c1-83ce-4061-9baa-351e27b30b09"). InnerVolumeSpecName "kube-api-access-8j585". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.039014 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60e2b0c1-83ce-4061-9baa-351e27b30b09" (UID: "60e2b0c1-83ce-4061-9baa-351e27b30b09"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.040751 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-config" (OuterVolumeSpecName: "config") pod "60e2b0c1-83ce-4061-9baa-351e27b30b09" (UID: "60e2b0c1-83ce-4061-9baa-351e27b30b09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.105314 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j585\" (UniqueName: \"kubernetes.io/projected/60e2b0c1-83ce-4061-9baa-351e27b30b09-kube-api-access-8j585\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.105583 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.105597 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60e2b0c1-83ce-4061-9baa-351e27b30b09-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.624962 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" event={"ID":"60e2b0c1-83ce-4061-9baa-351e27b30b09","Type":"ContainerDied","Data":"7a8b8eac8db63dd351b0513b7831be6361336d3fd2202820d424bb74c726de7e"} Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.625023 4935 scope.go:117] "RemoveContainer" containerID="2b4f1aff70ad057031684c6d6c77872d598fe45472224d32a3ef39d7a4bd3297" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.625822 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96d5866c7-n6qwp" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.645129 4935 scope.go:117] "RemoveContainer" containerID="97d71a2bc65cbd504001d9e7d84a34319439faf6f51a9aacf005b96ab7e98856" Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.659577 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-n6qwp"] Oct 05 08:44:33 crc kubenswrapper[4935]: I1005 08:44:33.664731 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-n6qwp"] Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.628765 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 05 08:44:34 crc kubenswrapper[4935]: E1005 08:44:34.631574 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e2b0c1-83ce-4061-9baa-351e27b30b09" containerName="dnsmasq-dns" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.631612 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e2b0c1-83ce-4061-9baa-351e27b30b09" containerName="dnsmasq-dns" Oct 05 08:44:34 crc kubenswrapper[4935]: E1005 08:44:34.631649 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e2b0c1-83ce-4061-9baa-351e27b30b09" containerName="init" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.631666 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e2b0c1-83ce-4061-9baa-351e27b30b09" containerName="init" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.632098 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e2b0c1-83ce-4061-9baa-351e27b30b09" containerName="dnsmasq-dns" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.633805 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.641034 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.642425 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-vxzds" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.642781 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.661759 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.729980 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-scripts\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.730066 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbdhf\" (UniqueName: \"kubernetes.io/projected/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-kube-api-access-bbdhf\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.730137 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.730158 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-config\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.730184 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.787421 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60e2b0c1-83ce-4061-9baa-351e27b30b09" path="/var/lib/kubelet/pods/60e2b0c1-83ce-4061-9baa-351e27b30b09/volumes" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.832296 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-scripts\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.832526 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbdhf\" (UniqueName: \"kubernetes.io/projected/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-kube-api-access-bbdhf\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.832706 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.832749 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-config\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.832833 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.833360 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.833671 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-scripts\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.833680 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-config\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.842199 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.858881 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbdhf\" (UniqueName: \"kubernetes.io/projected/ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5-kube-api-access-bbdhf\") pod \"ovn-northd-0\" (UID: \"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5\") " pod="openstack/ovn-northd-0" Oct 05 08:44:34 crc kubenswrapper[4935]: I1005 08:44:34.986926 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 08:44:35 crc kubenswrapper[4935]: I1005 08:44:35.462641 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 08:44:35 crc kubenswrapper[4935]: W1005 08:44:35.468087 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef6b6b47_12d1_4e7a_9f0c_ab654c363fd5.slice/crio-c614a7876e35a07bb8d90e659392e9b74d6e24b0ef763e38e2fe4738bdb90335 WatchSource:0}: Error finding container c614a7876e35a07bb8d90e659392e9b74d6e24b0ef763e38e2fe4738bdb90335: Status 404 returned error can't find the container with id c614a7876e35a07bb8d90e659392e9b74d6e24b0ef763e38e2fe4738bdb90335 Oct 05 08:44:35 crc kubenswrapper[4935]: I1005 08:44:35.656067 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5","Type":"ContainerStarted","Data":"c614a7876e35a07bb8d90e659392e9b74d6e24b0ef763e38e2fe4738bdb90335"} Oct 05 08:44:36 crc kubenswrapper[4935]: I1005 08:44:36.671281 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5","Type":"ContainerStarted","Data":"8ea8936ebfc729fe2eb96db1d2bab814eb9668e332dcc37527f97b7b94eca7ea"} Oct 05 08:44:36 crc kubenswrapper[4935]: I1005 08:44:36.671836 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5","Type":"ContainerStarted","Data":"3004525aa583da443ee4c62e3c6b4b855edcd390c8515295e73579b2afeb00d9"} Oct 05 08:44:36 crc kubenswrapper[4935]: I1005 08:44:36.672247 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 05 08:44:36 crc kubenswrapper[4935]: I1005 08:44:36.711919 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.083758551 podStartE2EDuration="2.71185654s" podCreationTimestamp="2025-10-05 08:44:34 +0000 UTC" firstStartedPulling="2025-10-05 08:44:35.471191185 +0000 UTC m=+6709.353817645" lastFinishedPulling="2025-10-05 08:44:36.099289174 +0000 UTC m=+6709.981915634" observedRunningTime="2025-10-05 08:44:36.697992381 +0000 UTC m=+6710.580618881" watchObservedRunningTime="2025-10-05 08:44:36.71185654 +0000 UTC m=+6710.594483020" Oct 05 08:44:43 crc kubenswrapper[4935]: I1005 08:44:43.333148 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ts4mj"] Oct 05 08:44:43 crc kubenswrapper[4935]: I1005 08:44:43.334576 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ts4mj" Oct 05 08:44:43 crc kubenswrapper[4935]: I1005 08:44:43.349295 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ts4mj"] Oct 05 08:44:43 crc kubenswrapper[4935]: I1005 08:44:43.411676 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjwp4\" (UniqueName: \"kubernetes.io/projected/95f38b74-c71b-4a9c-8ac7-4bb13037bc07-kube-api-access-hjwp4\") pod \"keystone-db-create-ts4mj\" (UID: \"95f38b74-c71b-4a9c-8ac7-4bb13037bc07\") " pod="openstack/keystone-db-create-ts4mj" Oct 05 08:44:43 crc kubenswrapper[4935]: I1005 08:44:43.513399 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjwp4\" (UniqueName: \"kubernetes.io/projected/95f38b74-c71b-4a9c-8ac7-4bb13037bc07-kube-api-access-hjwp4\") pod \"keystone-db-create-ts4mj\" (UID: \"95f38b74-c71b-4a9c-8ac7-4bb13037bc07\") " pod="openstack/keystone-db-create-ts4mj" Oct 05 08:44:43 crc kubenswrapper[4935]: I1005 08:44:43.544668 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjwp4\" (UniqueName: \"kubernetes.io/projected/95f38b74-c71b-4a9c-8ac7-4bb13037bc07-kube-api-access-hjwp4\") pod \"keystone-db-create-ts4mj\" (UID: \"95f38b74-c71b-4a9c-8ac7-4bb13037bc07\") " pod="openstack/keystone-db-create-ts4mj" Oct 05 08:44:43 crc kubenswrapper[4935]: I1005 08:44:43.698583 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ts4mj" Oct 05 08:44:44 crc kubenswrapper[4935]: I1005 08:44:44.218603 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ts4mj"] Oct 05 08:44:44 crc kubenswrapper[4935]: I1005 08:44:44.757020 4935 generic.go:334] "Generic (PLEG): container finished" podID="95f38b74-c71b-4a9c-8ac7-4bb13037bc07" containerID="3e58520a17cf1fe6c800531870e7f62071ac707bde020a522d6323c2b57f29ee" exitCode=0 Oct 05 08:44:44 crc kubenswrapper[4935]: I1005 08:44:44.757104 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ts4mj" event={"ID":"95f38b74-c71b-4a9c-8ac7-4bb13037bc07","Type":"ContainerDied","Data":"3e58520a17cf1fe6c800531870e7f62071ac707bde020a522d6323c2b57f29ee"} Oct 05 08:44:44 crc kubenswrapper[4935]: I1005 08:44:44.757303 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ts4mj" event={"ID":"95f38b74-c71b-4a9c-8ac7-4bb13037bc07","Type":"ContainerStarted","Data":"76aed09c9e3a3e986b8fca55dced2a75f6e64acba1681a6fc0041641e748d61a"} Oct 05 08:44:46 crc kubenswrapper[4935]: I1005 08:44:46.127787 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ts4mj" Oct 05 08:44:46 crc kubenswrapper[4935]: I1005 08:44:46.263274 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjwp4\" (UniqueName: \"kubernetes.io/projected/95f38b74-c71b-4a9c-8ac7-4bb13037bc07-kube-api-access-hjwp4\") pod \"95f38b74-c71b-4a9c-8ac7-4bb13037bc07\" (UID: \"95f38b74-c71b-4a9c-8ac7-4bb13037bc07\") " Oct 05 08:44:46 crc kubenswrapper[4935]: I1005 08:44:46.276592 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f38b74-c71b-4a9c-8ac7-4bb13037bc07-kube-api-access-hjwp4" (OuterVolumeSpecName: "kube-api-access-hjwp4") pod "95f38b74-c71b-4a9c-8ac7-4bb13037bc07" (UID: "95f38b74-c71b-4a9c-8ac7-4bb13037bc07"). InnerVolumeSpecName "kube-api-access-hjwp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:44:46 crc kubenswrapper[4935]: I1005 08:44:46.365067 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjwp4\" (UniqueName: \"kubernetes.io/projected/95f38b74-c71b-4a9c-8ac7-4bb13037bc07-kube-api-access-hjwp4\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:46 crc kubenswrapper[4935]: I1005 08:44:46.788786 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ts4mj" Oct 05 08:44:46 crc kubenswrapper[4935]: I1005 08:44:46.794011 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ts4mj" event={"ID":"95f38b74-c71b-4a9c-8ac7-4bb13037bc07","Type":"ContainerDied","Data":"76aed09c9e3a3e986b8fca55dced2a75f6e64acba1681a6fc0041641e748d61a"} Oct 05 08:44:46 crc kubenswrapper[4935]: I1005 08:44:46.794058 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76aed09c9e3a3e986b8fca55dced2a75f6e64acba1681a6fc0041641e748d61a" Oct 05 08:44:50 crc kubenswrapper[4935]: I1005 08:44:50.092213 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.331374 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2441-account-create-c9nbz"] Oct 05 08:44:53 crc kubenswrapper[4935]: E1005 08:44:53.332197 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f38b74-c71b-4a9c-8ac7-4bb13037bc07" containerName="mariadb-database-create" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.332210 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f38b74-c71b-4a9c-8ac7-4bb13037bc07" containerName="mariadb-database-create" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.332348 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f38b74-c71b-4a9c-8ac7-4bb13037bc07" containerName="mariadb-database-create" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.332914 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2441-account-create-c9nbz" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.335045 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.355675 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2441-account-create-c9nbz"] Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.390881 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwsxt\" (UniqueName: \"kubernetes.io/projected/23135322-e76e-457d-82d0-10463fefd374-kube-api-access-zwsxt\") pod \"keystone-2441-account-create-c9nbz\" (UID: \"23135322-e76e-457d-82d0-10463fefd374\") " pod="openstack/keystone-2441-account-create-c9nbz" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.493295 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwsxt\" (UniqueName: \"kubernetes.io/projected/23135322-e76e-457d-82d0-10463fefd374-kube-api-access-zwsxt\") pod \"keystone-2441-account-create-c9nbz\" (UID: \"23135322-e76e-457d-82d0-10463fefd374\") " pod="openstack/keystone-2441-account-create-c9nbz" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.517716 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwsxt\" (UniqueName: \"kubernetes.io/projected/23135322-e76e-457d-82d0-10463fefd374-kube-api-access-zwsxt\") pod \"keystone-2441-account-create-c9nbz\" (UID: \"23135322-e76e-457d-82d0-10463fefd374\") " pod="openstack/keystone-2441-account-create-c9nbz" Oct 05 08:44:53 crc kubenswrapper[4935]: I1005 08:44:53.654979 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2441-account-create-c9nbz" Oct 05 08:44:54 crc kubenswrapper[4935]: I1005 08:44:54.145620 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2441-account-create-c9nbz"] Oct 05 08:44:54 crc kubenswrapper[4935]: W1005 08:44:54.149786 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23135322_e76e_457d_82d0_10463fefd374.slice/crio-7e07fe1c09a502e531aa8d17d7119cffdbfa02dbf1fa4daa402924c84740f290 WatchSource:0}: Error finding container 7e07fe1c09a502e531aa8d17d7119cffdbfa02dbf1fa4daa402924c84740f290: Status 404 returned error can't find the container with id 7e07fe1c09a502e531aa8d17d7119cffdbfa02dbf1fa4daa402924c84740f290 Oct 05 08:44:54 crc kubenswrapper[4935]: I1005 08:44:54.871041 4935 generic.go:334] "Generic (PLEG): container finished" podID="23135322-e76e-457d-82d0-10463fefd374" containerID="759caccedc95fc23d130fab32680183f1cea064f59ff7f38b6b83195c0cd4a22" exitCode=0 Oct 05 08:44:54 crc kubenswrapper[4935]: I1005 08:44:54.871120 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2441-account-create-c9nbz" event={"ID":"23135322-e76e-457d-82d0-10463fefd374","Type":"ContainerDied","Data":"759caccedc95fc23d130fab32680183f1cea064f59ff7f38b6b83195c0cd4a22"} Oct 05 08:44:54 crc kubenswrapper[4935]: I1005 08:44:54.871169 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2441-account-create-c9nbz" event={"ID":"23135322-e76e-457d-82d0-10463fefd374","Type":"ContainerStarted","Data":"7e07fe1c09a502e531aa8d17d7119cffdbfa02dbf1fa4daa402924c84740f290"} Oct 05 08:44:56 crc kubenswrapper[4935]: I1005 08:44:56.205102 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2441-account-create-c9nbz" Oct 05 08:44:56 crc kubenswrapper[4935]: I1005 08:44:56.341767 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwsxt\" (UniqueName: \"kubernetes.io/projected/23135322-e76e-457d-82d0-10463fefd374-kube-api-access-zwsxt\") pod \"23135322-e76e-457d-82d0-10463fefd374\" (UID: \"23135322-e76e-457d-82d0-10463fefd374\") " Oct 05 08:44:56 crc kubenswrapper[4935]: I1005 08:44:56.354187 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23135322-e76e-457d-82d0-10463fefd374-kube-api-access-zwsxt" (OuterVolumeSpecName: "kube-api-access-zwsxt") pod "23135322-e76e-457d-82d0-10463fefd374" (UID: "23135322-e76e-457d-82d0-10463fefd374"). InnerVolumeSpecName "kube-api-access-zwsxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:44:56 crc kubenswrapper[4935]: I1005 08:44:56.444067 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwsxt\" (UniqueName: \"kubernetes.io/projected/23135322-e76e-457d-82d0-10463fefd374-kube-api-access-zwsxt\") on node \"crc\" DevicePath \"\"" Oct 05 08:44:56 crc kubenswrapper[4935]: I1005 08:44:56.888160 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2441-account-create-c9nbz" event={"ID":"23135322-e76e-457d-82d0-10463fefd374","Type":"ContainerDied","Data":"7e07fe1c09a502e531aa8d17d7119cffdbfa02dbf1fa4daa402924c84740f290"} Oct 05 08:44:56 crc kubenswrapper[4935]: I1005 08:44:56.888489 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e07fe1c09a502e531aa8d17d7119cffdbfa02dbf1fa4daa402924c84740f290" Oct 05 08:44:56 crc kubenswrapper[4935]: I1005 08:44:56.888221 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2441-account-create-c9nbz" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.726334 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-g7kbv"] Oct 05 08:44:58 crc kubenswrapper[4935]: E1005 08:44:58.726923 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23135322-e76e-457d-82d0-10463fefd374" containerName="mariadb-account-create" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.726942 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="23135322-e76e-457d-82d0-10463fefd374" containerName="mariadb-account-create" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.727187 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="23135322-e76e-457d-82d0-10463fefd374" containerName="mariadb-account-create" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.728372 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.732214 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-46sgw" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.732739 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.732937 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.734483 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.742233 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-g7kbv"] Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.784240 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-config-data\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.784553 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-combined-ca-bundle\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.784822 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qkdc\" (UniqueName: \"kubernetes.io/projected/308d5101-f413-4fd3-b5fa-90450a246f5f-kube-api-access-9qkdc\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.887510 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-config-data\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.887746 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-combined-ca-bundle\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.887793 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qkdc\" (UniqueName: \"kubernetes.io/projected/308d5101-f413-4fd3-b5fa-90450a246f5f-kube-api-access-9qkdc\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.894243 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-config-data\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.908770 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qkdc\" (UniqueName: \"kubernetes.io/projected/308d5101-f413-4fd3-b5fa-90450a246f5f-kube-api-access-9qkdc\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:58 crc kubenswrapper[4935]: I1005 08:44:58.915270 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-combined-ca-bundle\") pod \"keystone-db-sync-g7kbv\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:59 crc kubenswrapper[4935]: I1005 08:44:59.052564 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:44:59 crc kubenswrapper[4935]: I1005 08:44:59.595080 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-g7kbv"] Oct 05 08:44:59 crc kubenswrapper[4935]: I1005 08:44:59.913999 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g7kbv" event={"ID":"308d5101-f413-4fd3-b5fa-90450a246f5f","Type":"ContainerStarted","Data":"da231a658b40bcafd426d523ab965304bfc2511646dc646028884f439049891d"} Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.155809 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq"] Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.157584 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.160388 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.161786 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.167369 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq"] Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.209103 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-config-volume\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.209170 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x5mb\" (UniqueName: \"kubernetes.io/projected/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-kube-api-access-5x5mb\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.209207 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-secret-volume\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.311234 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-config-volume\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.312159 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-config-volume\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.312240 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x5mb\" (UniqueName: \"kubernetes.io/projected/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-kube-api-access-5x5mb\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.312525 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-secret-volume\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.320438 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-secret-volume\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.329158 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x5mb\" (UniqueName: \"kubernetes.io/projected/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-kube-api-access-5x5mb\") pod \"collect-profiles-29327565-sbhnq\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.491157 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:00 crc kubenswrapper[4935]: I1005 08:45:00.924435 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq"] Oct 05 08:45:00 crc kubenswrapper[4935]: W1005 08:45:00.924937 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ea58fa3_7301_4bbf_8bb1_38262bb5d5f1.slice/crio-8a78dfae8fe00ecf2b9bafbb167b48d0256186454c68772dbeb1881a4f3fe500 WatchSource:0}: Error finding container 8a78dfae8fe00ecf2b9bafbb167b48d0256186454c68772dbeb1881a4f3fe500: Status 404 returned error can't find the container with id 8a78dfae8fe00ecf2b9bafbb167b48d0256186454c68772dbeb1881a4f3fe500 Oct 05 08:45:01 crc kubenswrapper[4935]: I1005 08:45:01.931226 4935 generic.go:334] "Generic (PLEG): container finished" podID="2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1" containerID="3e46a7bdcd93059016a8b1ca22425a495c85a494178c9e958ed819f75ecb96c2" exitCode=0 Oct 05 08:45:01 crc kubenswrapper[4935]: I1005 08:45:01.931407 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" event={"ID":"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1","Type":"ContainerDied","Data":"3e46a7bdcd93059016a8b1ca22425a495c85a494178c9e958ed819f75ecb96c2"} Oct 05 08:45:01 crc kubenswrapper[4935]: I1005 08:45:01.931676 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" event={"ID":"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1","Type":"ContainerStarted","Data":"8a78dfae8fe00ecf2b9bafbb167b48d0256186454c68772dbeb1881a4f3fe500"} Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.508721 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.628222 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x5mb\" (UniqueName: \"kubernetes.io/projected/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-kube-api-access-5x5mb\") pod \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.628295 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-secret-volume\") pod \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.628432 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-config-volume\") pod \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\" (UID: \"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1\") " Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.629595 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-config-volume" (OuterVolumeSpecName: "config-volume") pod "2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1" (UID: "2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.632730 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1" (UID: "2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.632737 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-kube-api-access-5x5mb" (OuterVolumeSpecName: "kube-api-access-5x5mb") pod "2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1" (UID: "2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1"). InnerVolumeSpecName "kube-api-access-5x5mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.731322 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x5mb\" (UniqueName: \"kubernetes.io/projected/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-kube-api-access-5x5mb\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.731394 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.731409 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.966100 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" event={"ID":"2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1","Type":"ContainerDied","Data":"8a78dfae8fe00ecf2b9bafbb167b48d0256186454c68772dbeb1881a4f3fe500"} Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.966148 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.966177 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a78dfae8fe00ecf2b9bafbb167b48d0256186454c68772dbeb1881a4f3fe500" Oct 05 08:45:04 crc kubenswrapper[4935]: I1005 08:45:04.973116 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g7kbv" event={"ID":"308d5101-f413-4fd3-b5fa-90450a246f5f","Type":"ContainerStarted","Data":"8c00a3535c0c2b6e91f5b5aa2f990070f33798c6675ceb34c83653f2ce734e14"} Oct 05 08:45:05 crc kubenswrapper[4935]: I1005 08:45:05.006207 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-g7kbv" podStartSLOduration=2.111704187 podStartE2EDuration="7.006189201s" podCreationTimestamp="2025-10-05 08:44:58 +0000 UTC" firstStartedPulling="2025-10-05 08:44:59.60802337 +0000 UTC m=+6733.490649830" lastFinishedPulling="2025-10-05 08:45:04.502508374 +0000 UTC m=+6738.385134844" observedRunningTime="2025-10-05 08:45:05.000019367 +0000 UTC m=+6738.882645827" watchObservedRunningTime="2025-10-05 08:45:05.006189201 +0000 UTC m=+6738.888815681" Oct 05 08:45:05 crc kubenswrapper[4935]: I1005 08:45:05.611675 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b"] Oct 05 08:45:05 crc kubenswrapper[4935]: I1005 08:45:05.618554 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-s4x7b"] Oct 05 08:45:06 crc kubenswrapper[4935]: I1005 08:45:06.791702 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c163e86-9707-4230-8144-0016dd2b6c74" path="/var/lib/kubelet/pods/7c163e86-9707-4230-8144-0016dd2b6c74/volumes" Oct 05 08:45:06 crc kubenswrapper[4935]: I1005 08:45:06.994037 4935 generic.go:334] "Generic (PLEG): container finished" podID="308d5101-f413-4fd3-b5fa-90450a246f5f" containerID="8c00a3535c0c2b6e91f5b5aa2f990070f33798c6675ceb34c83653f2ce734e14" exitCode=0 Oct 05 08:45:06 crc kubenswrapper[4935]: I1005 08:45:06.994185 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g7kbv" event={"ID":"308d5101-f413-4fd3-b5fa-90450a246f5f","Type":"ContainerDied","Data":"8c00a3535c0c2b6e91f5b5aa2f990070f33798c6675ceb34c83653f2ce734e14"} Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.406744 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.499692 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qkdc\" (UniqueName: \"kubernetes.io/projected/308d5101-f413-4fd3-b5fa-90450a246f5f-kube-api-access-9qkdc\") pod \"308d5101-f413-4fd3-b5fa-90450a246f5f\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.499784 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-config-data\") pod \"308d5101-f413-4fd3-b5fa-90450a246f5f\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.500043 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-combined-ca-bundle\") pod \"308d5101-f413-4fd3-b5fa-90450a246f5f\" (UID: \"308d5101-f413-4fd3-b5fa-90450a246f5f\") " Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.508513 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308d5101-f413-4fd3-b5fa-90450a246f5f-kube-api-access-9qkdc" (OuterVolumeSpecName: "kube-api-access-9qkdc") pod "308d5101-f413-4fd3-b5fa-90450a246f5f" (UID: "308d5101-f413-4fd3-b5fa-90450a246f5f"). InnerVolumeSpecName "kube-api-access-9qkdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.523995 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "308d5101-f413-4fd3-b5fa-90450a246f5f" (UID: "308d5101-f413-4fd3-b5fa-90450a246f5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.550022 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-config-data" (OuterVolumeSpecName: "config-data") pod "308d5101-f413-4fd3-b5fa-90450a246f5f" (UID: "308d5101-f413-4fd3-b5fa-90450a246f5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.602379 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.602437 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qkdc\" (UniqueName: \"kubernetes.io/projected/308d5101-f413-4fd3-b5fa-90450a246f5f-kube-api-access-9qkdc\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:08 crc kubenswrapper[4935]: I1005 08:45:08.602450 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/308d5101-f413-4fd3-b5fa-90450a246f5f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.014217 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g7kbv" event={"ID":"308d5101-f413-4fd3-b5fa-90450a246f5f","Type":"ContainerDied","Data":"da231a658b40bcafd426d523ab965304bfc2511646dc646028884f439049891d"} Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.014258 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g7kbv" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.014275 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da231a658b40bcafd426d523ab965304bfc2511646dc646028884f439049891d" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.302861 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jv7s2"] Oct 05 08:45:09 crc kubenswrapper[4935]: E1005 08:45:09.303234 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="308d5101-f413-4fd3-b5fa-90450a246f5f" containerName="keystone-db-sync" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.303260 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="308d5101-f413-4fd3-b5fa-90450a246f5f" containerName="keystone-db-sync" Oct 05 08:45:09 crc kubenswrapper[4935]: E1005 08:45:09.303303 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1" containerName="collect-profiles" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.303312 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1" containerName="collect-profiles" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.303522 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="308d5101-f413-4fd3-b5fa-90450a246f5f" containerName="keystone-db-sync" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.303546 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1" containerName="collect-profiles" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.304172 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.308832 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.309265 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-46sgw" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.309998 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.310125 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.317500 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jv7s2"] Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.332905 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54c8c8bc55-xbt6n"] Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.334212 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.350544 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c8c8bc55-xbt6n"] Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.414994 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-combined-ca-bundle\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415049 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5dsh\" (UniqueName: \"kubernetes.io/projected/aa95a826-eeee-4268-a765-a755c5bc349c-kube-api-access-p5dsh\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415081 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-config-data\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415141 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-config\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415165 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-nb\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415195 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-scripts\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415254 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-credential-keys\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415280 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-dns-svc\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415300 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpjgv\" (UniqueName: \"kubernetes.io/projected/aef5cf15-d277-41a8-b6f8-a5016791e473-kube-api-access-dpjgv\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415336 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-fernet-keys\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.415372 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-sb\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.517346 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-config\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.518034 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-nb\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.518190 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-scripts\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.518332 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-credential-keys\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.518437 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-dns-svc\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.518520 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-config\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.518617 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpjgv\" (UniqueName: \"kubernetes.io/projected/aef5cf15-d277-41a8-b6f8-a5016791e473-kube-api-access-dpjgv\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.518737 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-fernet-keys\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.518861 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-sb\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.519031 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-combined-ca-bundle\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.519138 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5dsh\" (UniqueName: \"kubernetes.io/projected/aa95a826-eeee-4268-a765-a755c5bc349c-kube-api-access-p5dsh\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.519245 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-config-data\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.519678 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-nb\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.520980 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-dns-svc\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.521216 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-sb\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.521490 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-scripts\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.522296 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-fernet-keys\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.523206 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-credential-keys\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.523517 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-combined-ca-bundle\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.524111 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-config-data\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.540481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpjgv\" (UniqueName: \"kubernetes.io/projected/aef5cf15-d277-41a8-b6f8-a5016791e473-kube-api-access-dpjgv\") pod \"keystone-bootstrap-jv7s2\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.540779 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5dsh\" (UniqueName: \"kubernetes.io/projected/aa95a826-eeee-4268-a765-a755c5bc349c-kube-api-access-p5dsh\") pod \"dnsmasq-dns-54c8c8bc55-xbt6n\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.621970 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:09 crc kubenswrapper[4935]: I1005 08:45:09.654880 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:10 crc kubenswrapper[4935]: I1005 08:45:10.119917 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jv7s2"] Oct 05 08:45:10 crc kubenswrapper[4935]: W1005 08:45:10.154539 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaef5cf15_d277_41a8_b6f8_a5016791e473.slice/crio-9f612f072992987e29f6ff70c4829fc4125d36a1ed375fc057c64f1fa68206e1 WatchSource:0}: Error finding container 9f612f072992987e29f6ff70c4829fc4125d36a1ed375fc057c64f1fa68206e1: Status 404 returned error can't find the container with id 9f612f072992987e29f6ff70c4829fc4125d36a1ed375fc057c64f1fa68206e1 Oct 05 08:45:10 crc kubenswrapper[4935]: I1005 08:45:10.203429 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c8c8bc55-xbt6n"] Oct 05 08:45:10 crc kubenswrapper[4935]: W1005 08:45:10.210005 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa95a826_eeee_4268_a765_a755c5bc349c.slice/crio-f6bfb341599b9ee34ae0ee4b171607edf4f6b5d127772873aa49be03ff0a6ff2 WatchSource:0}: Error finding container f6bfb341599b9ee34ae0ee4b171607edf4f6b5d127772873aa49be03ff0a6ff2: Status 404 returned error can't find the container with id f6bfb341599b9ee34ae0ee4b171607edf4f6b5d127772873aa49be03ff0a6ff2 Oct 05 08:45:11 crc kubenswrapper[4935]: I1005 08:45:11.055766 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jv7s2" event={"ID":"aef5cf15-d277-41a8-b6f8-a5016791e473","Type":"ContainerStarted","Data":"99fb81a8f2a8422521134d63449b963210936354f48562b50f7605c4389011fa"} Oct 05 08:45:11 crc kubenswrapper[4935]: I1005 08:45:11.056097 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jv7s2" event={"ID":"aef5cf15-d277-41a8-b6f8-a5016791e473","Type":"ContainerStarted","Data":"9f612f072992987e29f6ff70c4829fc4125d36a1ed375fc057c64f1fa68206e1"} Oct 05 08:45:11 crc kubenswrapper[4935]: I1005 08:45:11.059524 4935 generic.go:334] "Generic (PLEG): container finished" podID="aa95a826-eeee-4268-a765-a755c5bc349c" containerID="99f0c5c556ddcb36370a192dbbcb4368af41583ae8347670d9a10607598d7bc4" exitCode=0 Oct 05 08:45:11 crc kubenswrapper[4935]: I1005 08:45:11.059557 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" event={"ID":"aa95a826-eeee-4268-a765-a755c5bc349c","Type":"ContainerDied","Data":"99f0c5c556ddcb36370a192dbbcb4368af41583ae8347670d9a10607598d7bc4"} Oct 05 08:45:11 crc kubenswrapper[4935]: I1005 08:45:11.059574 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" event={"ID":"aa95a826-eeee-4268-a765-a755c5bc349c","Type":"ContainerStarted","Data":"f6bfb341599b9ee34ae0ee4b171607edf4f6b5d127772873aa49be03ff0a6ff2"} Oct 05 08:45:11 crc kubenswrapper[4935]: I1005 08:45:11.084722 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jv7s2" podStartSLOduration=2.084704512 podStartE2EDuration="2.084704512s" podCreationTimestamp="2025-10-05 08:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:45:11.084435654 +0000 UTC m=+6744.967062164" watchObservedRunningTime="2025-10-05 08:45:11.084704512 +0000 UTC m=+6744.967330982" Oct 05 08:45:12 crc kubenswrapper[4935]: I1005 08:45:12.081186 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" event={"ID":"aa95a826-eeee-4268-a765-a755c5bc349c","Type":"ContainerStarted","Data":"1d57c2318d21c9fcf6ce9760e9eaeee91bb3b8ee734fecfc5623918793bce431"} Oct 05 08:45:12 crc kubenswrapper[4935]: I1005 08:45:12.086468 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:12 crc kubenswrapper[4935]: I1005 08:45:12.119823 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" podStartSLOduration=3.119803495 podStartE2EDuration="3.119803495s" podCreationTimestamp="2025-10-05 08:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:45:12.113837006 +0000 UTC m=+6745.996463506" watchObservedRunningTime="2025-10-05 08:45:12.119803495 +0000 UTC m=+6746.002429965" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.359941 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x79wz"] Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.362398 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.370301 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x79wz"] Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.412223 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc8qj\" (UniqueName: \"kubernetes.io/projected/321c5eb7-cd7f-405d-84be-84e654a57c85-kube-api-access-qc8qj\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.412314 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-catalog-content\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.412476 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-utilities\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.515258 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-utilities\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.515419 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc8qj\" (UniqueName: \"kubernetes.io/projected/321c5eb7-cd7f-405d-84be-84e654a57c85-kube-api-access-qc8qj\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.515470 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-catalog-content\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.515844 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-utilities\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.516203 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-catalog-content\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.536195 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc8qj\" (UniqueName: \"kubernetes.io/projected/321c5eb7-cd7f-405d-84be-84e654a57c85-kube-api-access-qc8qj\") pod \"community-operators-x79wz\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:13 crc kubenswrapper[4935]: I1005 08:45:13.709620 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:14 crc kubenswrapper[4935]: I1005 08:45:14.098181 4935 generic.go:334] "Generic (PLEG): container finished" podID="aef5cf15-d277-41a8-b6f8-a5016791e473" containerID="99fb81a8f2a8422521134d63449b963210936354f48562b50f7605c4389011fa" exitCode=0 Oct 05 08:45:14 crc kubenswrapper[4935]: I1005 08:45:14.098475 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jv7s2" event={"ID":"aef5cf15-d277-41a8-b6f8-a5016791e473","Type":"ContainerDied","Data":"99fb81a8f2a8422521134d63449b963210936354f48562b50f7605c4389011fa"} Oct 05 08:45:14 crc kubenswrapper[4935]: I1005 08:45:14.215624 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x79wz"] Oct 05 08:45:14 crc kubenswrapper[4935]: I1005 08:45:14.293065 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:45:14 crc kubenswrapper[4935]: I1005 08:45:14.293131 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.109362 4935 generic.go:334] "Generic (PLEG): container finished" podID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerID="39d994f461a32cf50f0932edb6149c14f8cc8201b0e968da6a9ea2dfac8f073d" exitCode=0 Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.109478 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79wz" event={"ID":"321c5eb7-cd7f-405d-84be-84e654a57c85","Type":"ContainerDied","Data":"39d994f461a32cf50f0932edb6149c14f8cc8201b0e968da6a9ea2dfac8f073d"} Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.109726 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79wz" event={"ID":"321c5eb7-cd7f-405d-84be-84e654a57c85","Type":"ContainerStarted","Data":"883fe0bd026a5d74e036c206912cefac6c0e9cf57240ddb678a16da3f4d7756b"} Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.450708 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.584141 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-scripts\") pod \"aef5cf15-d277-41a8-b6f8-a5016791e473\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.584252 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-fernet-keys\") pod \"aef5cf15-d277-41a8-b6f8-a5016791e473\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.584354 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-credential-keys\") pod \"aef5cf15-d277-41a8-b6f8-a5016791e473\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.584381 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-config-data\") pod \"aef5cf15-d277-41a8-b6f8-a5016791e473\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.584418 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpjgv\" (UniqueName: \"kubernetes.io/projected/aef5cf15-d277-41a8-b6f8-a5016791e473-kube-api-access-dpjgv\") pod \"aef5cf15-d277-41a8-b6f8-a5016791e473\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.584471 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-combined-ca-bundle\") pod \"aef5cf15-d277-41a8-b6f8-a5016791e473\" (UID: \"aef5cf15-d277-41a8-b6f8-a5016791e473\") " Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.589764 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef5cf15-d277-41a8-b6f8-a5016791e473-kube-api-access-dpjgv" (OuterVolumeSpecName: "kube-api-access-dpjgv") pod "aef5cf15-d277-41a8-b6f8-a5016791e473" (UID: "aef5cf15-d277-41a8-b6f8-a5016791e473"). InnerVolumeSpecName "kube-api-access-dpjgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.593527 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-scripts" (OuterVolumeSpecName: "scripts") pod "aef5cf15-d277-41a8-b6f8-a5016791e473" (UID: "aef5cf15-d277-41a8-b6f8-a5016791e473"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.603093 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "aef5cf15-d277-41a8-b6f8-a5016791e473" (UID: "aef5cf15-d277-41a8-b6f8-a5016791e473"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.603978 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "aef5cf15-d277-41a8-b6f8-a5016791e473" (UID: "aef5cf15-d277-41a8-b6f8-a5016791e473"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.619766 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-config-data" (OuterVolumeSpecName: "config-data") pod "aef5cf15-d277-41a8-b6f8-a5016791e473" (UID: "aef5cf15-d277-41a8-b6f8-a5016791e473"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.621364 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aef5cf15-d277-41a8-b6f8-a5016791e473" (UID: "aef5cf15-d277-41a8-b6f8-a5016791e473"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.686270 4935 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.686306 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.686317 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpjgv\" (UniqueName: \"kubernetes.io/projected/aef5cf15-d277-41a8-b6f8-a5016791e473-kube-api-access-dpjgv\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.686341 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.686350 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:15 crc kubenswrapper[4935]: I1005 08:45:15.686358 4935 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aef5cf15-d277-41a8-b6f8-a5016791e473-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.120770 4935 generic.go:334] "Generic (PLEG): container finished" podID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerID="4042722c3c48e582f0e47e231d395266889a4315ca99fc8831403ca27161085c" exitCode=0 Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.120874 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79wz" event={"ID":"321c5eb7-cd7f-405d-84be-84e654a57c85","Type":"ContainerDied","Data":"4042722c3c48e582f0e47e231d395266889a4315ca99fc8831403ca27161085c"} Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.123778 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jv7s2" event={"ID":"aef5cf15-d277-41a8-b6f8-a5016791e473","Type":"ContainerDied","Data":"9f612f072992987e29f6ff70c4829fc4125d36a1ed375fc057c64f1fa68206e1"} Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.123928 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f612f072992987e29f6ff70c4829fc4125d36a1ed375fc057c64f1fa68206e1" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.123866 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jv7s2" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.192777 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jv7s2"] Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.198254 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jv7s2"] Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.299982 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-t7nvk"] Oct 05 08:45:16 crc kubenswrapper[4935]: E1005 08:45:16.300365 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef5cf15-d277-41a8-b6f8-a5016791e473" containerName="keystone-bootstrap" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.300385 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef5cf15-d277-41a8-b6f8-a5016791e473" containerName="keystone-bootstrap" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.300598 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef5cf15-d277-41a8-b6f8-a5016791e473" containerName="keystone-bootstrap" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.301211 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.312591 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.312925 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.313090 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-46sgw" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.313321 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.316972 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t7nvk"] Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.399949 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-credential-keys\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.400031 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-fernet-keys\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.400057 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-config-data\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.400117 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-combined-ca-bundle\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.400169 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-scripts\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.400271 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf9z4\" (UniqueName: \"kubernetes.io/projected/37b4f125-73af-4d84-bc8e-a11257c7df45-kube-api-access-bf9z4\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.501669 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-credential-keys\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.501728 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-fernet-keys\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.501752 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-config-data\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.501771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-combined-ca-bundle\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.501793 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-scripts\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.501838 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf9z4\" (UniqueName: \"kubernetes.io/projected/37b4f125-73af-4d84-bc8e-a11257c7df45-kube-api-access-bf9z4\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.508443 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-credential-keys\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.510052 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-combined-ca-bundle\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.513659 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-scripts\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.514791 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-config-data\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.526481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-fernet-keys\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.528482 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf9z4\" (UniqueName: \"kubernetes.io/projected/37b4f125-73af-4d84-bc8e-a11257c7df45-kube-api-access-bf9z4\") pod \"keystone-bootstrap-t7nvk\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.626103 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:16 crc kubenswrapper[4935]: I1005 08:45:16.839073 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef5cf15-d277-41a8-b6f8-a5016791e473" path="/var/lib/kubelet/pods/aef5cf15-d277-41a8-b6f8-a5016791e473/volumes" Oct 05 08:45:17 crc kubenswrapper[4935]: I1005 08:45:17.125213 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t7nvk"] Oct 05 08:45:17 crc kubenswrapper[4935]: I1005 08:45:17.133040 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79wz" event={"ID":"321c5eb7-cd7f-405d-84be-84e654a57c85","Type":"ContainerStarted","Data":"581f062ac3b0d74adf16872bb5080539ac36ecd2e35eb50f0df16615faa0bd9c"} Oct 05 08:45:17 crc kubenswrapper[4935]: W1005 08:45:17.135190 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37b4f125_73af_4d84_bc8e_a11257c7df45.slice/crio-3a1312eb524d6df131d14aa2c00f9cc9156e8e0b52de4c6582c5a3270aab0c16 WatchSource:0}: Error finding container 3a1312eb524d6df131d14aa2c00f9cc9156e8e0b52de4c6582c5a3270aab0c16: Status 404 returned error can't find the container with id 3a1312eb524d6df131d14aa2c00f9cc9156e8e0b52de4c6582c5a3270aab0c16 Oct 05 08:45:18 crc kubenswrapper[4935]: I1005 08:45:18.140669 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t7nvk" event={"ID":"37b4f125-73af-4d84-bc8e-a11257c7df45","Type":"ContainerStarted","Data":"1c97190c183f6cd258b44028c533845f0d0cba070667fbd128b80f06aa6301bf"} Oct 05 08:45:18 crc kubenswrapper[4935]: I1005 08:45:18.141004 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t7nvk" event={"ID":"37b4f125-73af-4d84-bc8e-a11257c7df45","Type":"ContainerStarted","Data":"3a1312eb524d6df131d14aa2c00f9cc9156e8e0b52de4c6582c5a3270aab0c16"} Oct 05 08:45:18 crc kubenswrapper[4935]: I1005 08:45:18.159477 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-t7nvk" podStartSLOduration=2.15946115 podStartE2EDuration="2.15946115s" podCreationTimestamp="2025-10-05 08:45:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:45:18.157464867 +0000 UTC m=+6752.040091327" watchObservedRunningTime="2025-10-05 08:45:18.15946115 +0000 UTC m=+6752.042087610" Oct 05 08:45:18 crc kubenswrapper[4935]: I1005 08:45:18.168244 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x79wz" podStartSLOduration=3.780030682 podStartE2EDuration="5.168223404s" podCreationTimestamp="2025-10-05 08:45:13 +0000 UTC" firstStartedPulling="2025-10-05 08:45:15.111834567 +0000 UTC m=+6748.994461037" lastFinishedPulling="2025-10-05 08:45:16.500027299 +0000 UTC m=+6750.382653759" observedRunningTime="2025-10-05 08:45:17.156492313 +0000 UTC m=+6751.039118793" watchObservedRunningTime="2025-10-05 08:45:18.168223404 +0000 UTC m=+6752.050849864" Oct 05 08:45:19 crc kubenswrapper[4935]: I1005 08:45:19.656050 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:45:19 crc kubenswrapper[4935]: I1005 08:45:19.733578 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b486dfb5c-xkdlz"] Oct 05 08:45:19 crc kubenswrapper[4935]: I1005 08:45:19.734836 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" podUID="957fa2eb-5dc0-40c2-9252-2161d11ce895" containerName="dnsmasq-dns" containerID="cri-o://ae1dab29c65968860231c27e67156bdb9501a55e08cd293fa1e400be513d97f3" gracePeriod=10 Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.168279 4935 generic.go:334] "Generic (PLEG): container finished" podID="37b4f125-73af-4d84-bc8e-a11257c7df45" containerID="1c97190c183f6cd258b44028c533845f0d0cba070667fbd128b80f06aa6301bf" exitCode=0 Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.168863 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t7nvk" event={"ID":"37b4f125-73af-4d84-bc8e-a11257c7df45","Type":"ContainerDied","Data":"1c97190c183f6cd258b44028c533845f0d0cba070667fbd128b80f06aa6301bf"} Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.175699 4935 generic.go:334] "Generic (PLEG): container finished" podID="957fa2eb-5dc0-40c2-9252-2161d11ce895" containerID="ae1dab29c65968860231c27e67156bdb9501a55e08cd293fa1e400be513d97f3" exitCode=0 Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.175774 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" event={"ID":"957fa2eb-5dc0-40c2-9252-2161d11ce895","Type":"ContainerDied","Data":"ae1dab29c65968860231c27e67156bdb9501a55e08cd293fa1e400be513d97f3"} Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.175831 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" event={"ID":"957fa2eb-5dc0-40c2-9252-2161d11ce895","Type":"ContainerDied","Data":"2bace3c8ad78200307dfa3f1baa907c6d7235e1c6483d46bb6dcc28b35d1192e"} Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.175844 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bace3c8ad78200307dfa3f1baa907c6d7235e1c6483d46bb6dcc28b35d1192e" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.225511 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.279633 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-sb\") pod \"957fa2eb-5dc0-40c2-9252-2161d11ce895\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.279697 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-config\") pod \"957fa2eb-5dc0-40c2-9252-2161d11ce895\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.279813 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-dns-svc\") pod \"957fa2eb-5dc0-40c2-9252-2161d11ce895\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.280720 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddpqd\" (UniqueName: \"kubernetes.io/projected/957fa2eb-5dc0-40c2-9252-2161d11ce895-kube-api-access-ddpqd\") pod \"957fa2eb-5dc0-40c2-9252-2161d11ce895\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.280813 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-nb\") pod \"957fa2eb-5dc0-40c2-9252-2161d11ce895\" (UID: \"957fa2eb-5dc0-40c2-9252-2161d11ce895\") " Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.290553 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/957fa2eb-5dc0-40c2-9252-2161d11ce895-kube-api-access-ddpqd" (OuterVolumeSpecName: "kube-api-access-ddpqd") pod "957fa2eb-5dc0-40c2-9252-2161d11ce895" (UID: "957fa2eb-5dc0-40c2-9252-2161d11ce895"). InnerVolumeSpecName "kube-api-access-ddpqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.322875 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "957fa2eb-5dc0-40c2-9252-2161d11ce895" (UID: "957fa2eb-5dc0-40c2-9252-2161d11ce895"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.322928 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "957fa2eb-5dc0-40c2-9252-2161d11ce895" (UID: "957fa2eb-5dc0-40c2-9252-2161d11ce895"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.323915 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "957fa2eb-5dc0-40c2-9252-2161d11ce895" (UID: "957fa2eb-5dc0-40c2-9252-2161d11ce895"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.324462 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-config" (OuterVolumeSpecName: "config") pod "957fa2eb-5dc0-40c2-9252-2161d11ce895" (UID: "957fa2eb-5dc0-40c2-9252-2161d11ce895"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.382960 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.382995 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.383005 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddpqd\" (UniqueName: \"kubernetes.io/projected/957fa2eb-5dc0-40c2-9252-2161d11ce895-kube-api-access-ddpqd\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.383014 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:20 crc kubenswrapper[4935]: I1005 08:45:20.383024 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/957fa2eb-5dc0-40c2-9252-2161d11ce895-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.182703 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b486dfb5c-xkdlz" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.217358 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b486dfb5c-xkdlz"] Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.226738 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b486dfb5c-xkdlz"] Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.519518 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.603506 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-scripts\") pod \"37b4f125-73af-4d84-bc8e-a11257c7df45\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.603635 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-fernet-keys\") pod \"37b4f125-73af-4d84-bc8e-a11257c7df45\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.603687 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-credential-keys\") pod \"37b4f125-73af-4d84-bc8e-a11257c7df45\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.603739 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-combined-ca-bundle\") pod \"37b4f125-73af-4d84-bc8e-a11257c7df45\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.603803 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf9z4\" (UniqueName: \"kubernetes.io/projected/37b4f125-73af-4d84-bc8e-a11257c7df45-kube-api-access-bf9z4\") pod \"37b4f125-73af-4d84-bc8e-a11257c7df45\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.603993 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-config-data\") pod \"37b4f125-73af-4d84-bc8e-a11257c7df45\" (UID: \"37b4f125-73af-4d84-bc8e-a11257c7df45\") " Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.612980 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "37b4f125-73af-4d84-bc8e-a11257c7df45" (UID: "37b4f125-73af-4d84-bc8e-a11257c7df45"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.614359 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-scripts" (OuterVolumeSpecName: "scripts") pod "37b4f125-73af-4d84-bc8e-a11257c7df45" (UID: "37b4f125-73af-4d84-bc8e-a11257c7df45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.616440 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b4f125-73af-4d84-bc8e-a11257c7df45-kube-api-access-bf9z4" (OuterVolumeSpecName: "kube-api-access-bf9z4") pod "37b4f125-73af-4d84-bc8e-a11257c7df45" (UID: "37b4f125-73af-4d84-bc8e-a11257c7df45"). InnerVolumeSpecName "kube-api-access-bf9z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.617144 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "37b4f125-73af-4d84-bc8e-a11257c7df45" (UID: "37b4f125-73af-4d84-bc8e-a11257c7df45"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.627549 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-config-data" (OuterVolumeSpecName: "config-data") pod "37b4f125-73af-4d84-bc8e-a11257c7df45" (UID: "37b4f125-73af-4d84-bc8e-a11257c7df45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.630044 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37b4f125-73af-4d84-bc8e-a11257c7df45" (UID: "37b4f125-73af-4d84-bc8e-a11257c7df45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.706266 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.706317 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.706334 4935 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.706346 4935 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.706358 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4f125-73af-4d84-bc8e-a11257c7df45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:21 crc kubenswrapper[4935]: I1005 08:45:21.706369 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf9z4\" (UniqueName: \"kubernetes.io/projected/37b4f125-73af-4d84-bc8e-a11257c7df45-kube-api-access-bf9z4\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.193448 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t7nvk" event={"ID":"37b4f125-73af-4d84-bc8e-a11257c7df45","Type":"ContainerDied","Data":"3a1312eb524d6df131d14aa2c00f9cc9156e8e0b52de4c6582c5a3270aab0c16"} Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.193499 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a1312eb524d6df131d14aa2c00f9cc9156e8e0b52de4c6582c5a3270aab0c16" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.194712 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t7nvk" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.258987 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f8f877d-mhddp"] Oct 05 08:45:22 crc kubenswrapper[4935]: E1005 08:45:22.259831 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957fa2eb-5dc0-40c2-9252-2161d11ce895" containerName="dnsmasq-dns" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.259854 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="957fa2eb-5dc0-40c2-9252-2161d11ce895" containerName="dnsmasq-dns" Oct 05 08:45:22 crc kubenswrapper[4935]: E1005 08:45:22.259865 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b4f125-73af-4d84-bc8e-a11257c7df45" containerName="keystone-bootstrap" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.259873 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b4f125-73af-4d84-bc8e-a11257c7df45" containerName="keystone-bootstrap" Oct 05 08:45:22 crc kubenswrapper[4935]: E1005 08:45:22.259910 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957fa2eb-5dc0-40c2-9252-2161d11ce895" containerName="init" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.259951 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="957fa2eb-5dc0-40c2-9252-2161d11ce895" containerName="init" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.260231 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b4f125-73af-4d84-bc8e-a11257c7df45" containerName="keystone-bootstrap" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.260288 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="957fa2eb-5dc0-40c2-9252-2161d11ce895" containerName="dnsmasq-dns" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.261034 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.263377 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.263718 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-46sgw" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.263716 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.263840 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.272037 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f8f877d-mhddp"] Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.316127 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-combined-ca-bundle\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.316395 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-scripts\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.316516 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-config-data\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.316683 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k6xl\" (UniqueName: \"kubernetes.io/projected/3c556c2f-6744-4691-b651-e5359f1c8d78-kube-api-access-4k6xl\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.316782 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-credential-keys\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.316817 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-fernet-keys\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.418534 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k6xl\" (UniqueName: \"kubernetes.io/projected/3c556c2f-6744-4691-b651-e5359f1c8d78-kube-api-access-4k6xl\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.418628 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-credential-keys\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.418662 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-fernet-keys\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.418717 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-combined-ca-bundle\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.418752 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-scripts\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.418779 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-config-data\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.424497 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-credential-keys\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.424688 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-combined-ca-bundle\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.424709 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-fernet-keys\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.424958 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-scripts\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.425835 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c556c2f-6744-4691-b651-e5359f1c8d78-config-data\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.435910 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k6xl\" (UniqueName: \"kubernetes.io/projected/3c556c2f-6744-4691-b651-e5359f1c8d78-kube-api-access-4k6xl\") pod \"keystone-f8f877d-mhddp\" (UID: \"3c556c2f-6744-4691-b651-e5359f1c8d78\") " pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.575749 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:22 crc kubenswrapper[4935]: I1005 08:45:22.786504 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="957fa2eb-5dc0-40c2-9252-2161d11ce895" path="/var/lib/kubelet/pods/957fa2eb-5dc0-40c2-9252-2161d11ce895/volumes" Oct 05 08:45:23 crc kubenswrapper[4935]: I1005 08:45:23.021693 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f8f877d-mhddp"] Oct 05 08:45:23 crc kubenswrapper[4935]: I1005 08:45:23.216543 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f8f877d-mhddp" event={"ID":"3c556c2f-6744-4691-b651-e5359f1c8d78","Type":"ContainerStarted","Data":"68b3df4ace51d5064727e595f9fa239c4fcb4be050dd9549a5b4c8a5e218a18c"} Oct 05 08:45:23 crc kubenswrapper[4935]: I1005 08:45:23.710202 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:23 crc kubenswrapper[4935]: I1005 08:45:23.710530 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:23 crc kubenswrapper[4935]: I1005 08:45:23.756112 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:24 crc kubenswrapper[4935]: I1005 08:45:24.227656 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f8f877d-mhddp" event={"ID":"3c556c2f-6744-4691-b651-e5359f1c8d78","Type":"ContainerStarted","Data":"555f552fb048374e0bf2341880a90f6eccc92ef2eba7430442edec3a34c2343e"} Oct 05 08:45:24 crc kubenswrapper[4935]: I1005 08:45:24.227736 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:24 crc kubenswrapper[4935]: I1005 08:45:24.309570 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:24 crc kubenswrapper[4935]: I1005 08:45:24.332339 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-f8f877d-mhddp" podStartSLOduration=2.332318584 podStartE2EDuration="2.332318584s" podCreationTimestamp="2025-10-05 08:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:45:24.251166724 +0000 UTC m=+6758.133793204" watchObservedRunningTime="2025-10-05 08:45:24.332318584 +0000 UTC m=+6758.214945054" Oct 05 08:45:24 crc kubenswrapper[4935]: I1005 08:45:24.357062 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x79wz"] Oct 05 08:45:26 crc kubenswrapper[4935]: I1005 08:45:26.246690 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x79wz" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerName="registry-server" containerID="cri-o://581f062ac3b0d74adf16872bb5080539ac36ecd2e35eb50f0df16615faa0bd9c" gracePeriod=2 Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.268021 4935 generic.go:334] "Generic (PLEG): container finished" podID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerID="581f062ac3b0d74adf16872bb5080539ac36ecd2e35eb50f0df16615faa0bd9c" exitCode=0 Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.268110 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79wz" event={"ID":"321c5eb7-cd7f-405d-84be-84e654a57c85","Type":"ContainerDied","Data":"581f062ac3b0d74adf16872bb5080539ac36ecd2e35eb50f0df16615faa0bd9c"} Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.268399 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x79wz" event={"ID":"321c5eb7-cd7f-405d-84be-84e654a57c85","Type":"ContainerDied","Data":"883fe0bd026a5d74e036c206912cefac6c0e9cf57240ddb678a16da3f4d7756b"} Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.268430 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="883fe0bd026a5d74e036c206912cefac6c0e9cf57240ddb678a16da3f4d7756b" Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.302534 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.409034 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc8qj\" (UniqueName: \"kubernetes.io/projected/321c5eb7-cd7f-405d-84be-84e654a57c85-kube-api-access-qc8qj\") pod \"321c5eb7-cd7f-405d-84be-84e654a57c85\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.409217 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-catalog-content\") pod \"321c5eb7-cd7f-405d-84be-84e654a57c85\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.409533 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-utilities\") pod \"321c5eb7-cd7f-405d-84be-84e654a57c85\" (UID: \"321c5eb7-cd7f-405d-84be-84e654a57c85\") " Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.411353 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-utilities" (OuterVolumeSpecName: "utilities") pod "321c5eb7-cd7f-405d-84be-84e654a57c85" (UID: "321c5eb7-cd7f-405d-84be-84e654a57c85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.414956 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/321c5eb7-cd7f-405d-84be-84e654a57c85-kube-api-access-qc8qj" (OuterVolumeSpecName: "kube-api-access-qc8qj") pod "321c5eb7-cd7f-405d-84be-84e654a57c85" (UID: "321c5eb7-cd7f-405d-84be-84e654a57c85"). InnerVolumeSpecName "kube-api-access-qc8qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.477645 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "321c5eb7-cd7f-405d-84be-84e654a57c85" (UID: "321c5eb7-cd7f-405d-84be-84e654a57c85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.511618 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc8qj\" (UniqueName: \"kubernetes.io/projected/321c5eb7-cd7f-405d-84be-84e654a57c85-kube-api-access-qc8qj\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.517511 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:27 crc kubenswrapper[4935]: I1005 08:45:27.517536 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/321c5eb7-cd7f-405d-84be-84e654a57c85-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:28 crc kubenswrapper[4935]: I1005 08:45:28.280276 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x79wz" Oct 05 08:45:28 crc kubenswrapper[4935]: I1005 08:45:28.318959 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x79wz"] Oct 05 08:45:28 crc kubenswrapper[4935]: I1005 08:45:28.323802 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x79wz"] Oct 05 08:45:28 crc kubenswrapper[4935]: I1005 08:45:28.796226 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" path="/var/lib/kubelet/pods/321c5eb7-cd7f-405d-84be-84e654a57c85/volumes" Oct 05 08:45:44 crc kubenswrapper[4935]: I1005 08:45:44.289936 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:45:44 crc kubenswrapper[4935]: I1005 08:45:44.290769 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:45:53 crc kubenswrapper[4935]: I1005 08:45:53.997647 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-f8f877d-mhddp" Oct 05 08:45:57 crc kubenswrapper[4935]: I1005 08:45:57.223477 4935 scope.go:117] "RemoveContainer" containerID="5ebe97d01e4970397aeb73ed4255c3c70f57a88ade2bb6699d24142822b5645b" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.377726 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 08:45:58 crc kubenswrapper[4935]: E1005 08:45:58.378267 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerName="extract-content" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.378292 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerName="extract-content" Oct 05 08:45:58 crc kubenswrapper[4935]: E1005 08:45:58.378354 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerName="extract-utilities" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.378367 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerName="extract-utilities" Oct 05 08:45:58 crc kubenswrapper[4935]: E1005 08:45:58.378392 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerName="registry-server" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.378405 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerName="registry-server" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.378718 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="321c5eb7-cd7f-405d-84be-84e654a57c85" containerName="registry-server" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.379656 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.381841 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.382297 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-npr8f" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.382491 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.384670 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.398321 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.405731 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 05 08:45:58 crc kubenswrapper[4935]: E1005 08:45:58.420211 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-zv9sw openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[kube-api-access-zv9sw openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="5e2c5abb-1271-4c6e-acbe-3c3137c78df0" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.438221 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.439975 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.445106 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5e2c5abb-1271-4c6e-acbe-3c3137c78df0" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.470041 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.578562 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.578630 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config-secret\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.578659 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcs8c\" (UniqueName: \"kubernetes.io/projected/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-kube-api-access-fcs8c\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.657808 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.665359 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5e2c5abb-1271-4c6e-acbe-3c3137c78df0" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.668186 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.671112 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5e2c5abb-1271-4c6e-acbe-3c3137c78df0" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.679989 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config-secret\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.680033 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcs8c\" (UniqueName: \"kubernetes.io/projected/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-kube-api-access-fcs8c\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.680133 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.682013 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.702670 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config-secret\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.722506 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcs8c\" (UniqueName: \"kubernetes.io/projected/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-kube-api-access-fcs8c\") pod \"openstackclient\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.766616 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:45:58 crc kubenswrapper[4935]: I1005 08:45:58.786441 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2c5abb-1271-4c6e-acbe-3c3137c78df0" path="/var/lib/kubelet/pods/5e2c5abb-1271-4c6e-acbe-3c3137c78df0/volumes" Oct 05 08:45:59 crc kubenswrapper[4935]: I1005 08:45:59.193794 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:45:59 crc kubenswrapper[4935]: W1005 08:45:59.198265 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5810bf2f_6c1e_465b_9ad4_a17ac78a0a56.slice/crio-75c3f2f03e9e7b51e46b97d923628f0732b0e5155dfb8c915c49f1b8be879531 WatchSource:0}: Error finding container 75c3f2f03e9e7b51e46b97d923628f0732b0e5155dfb8c915c49f1b8be879531: Status 404 returned error can't find the container with id 75c3f2f03e9e7b51e46b97d923628f0732b0e5155dfb8c915c49f1b8be879531 Oct 05 08:45:59 crc kubenswrapper[4935]: I1005 08:45:59.667295 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56","Type":"ContainerStarted","Data":"75c3f2f03e9e7b51e46b97d923628f0732b0e5155dfb8c915c49f1b8be879531"} Oct 05 08:45:59 crc kubenswrapper[4935]: I1005 08:45:59.667327 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:45:59 crc kubenswrapper[4935]: I1005 08:45:59.671106 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5e2c5abb-1271-4c6e-acbe-3c3137c78df0" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" Oct 05 08:45:59 crc kubenswrapper[4935]: I1005 08:45:59.678461 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5e2c5abb-1271-4c6e-acbe-3c3137c78df0" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" Oct 05 08:46:09 crc kubenswrapper[4935]: I1005 08:46:09.745752 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56","Type":"ContainerStarted","Data":"3ff8e08c666a64c87ec949ce6136a4e8df7ced86690ff5502b8fbdba5138fcb4"} Oct 05 08:46:09 crc kubenswrapper[4935]: I1005 08:46:09.763547 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.112277264 podStartE2EDuration="11.763529105s" podCreationTimestamp="2025-10-05 08:45:58 +0000 UTC" firstStartedPulling="2025-10-05 08:45:59.201458429 +0000 UTC m=+6793.084084889" lastFinishedPulling="2025-10-05 08:46:08.85271027 +0000 UTC m=+6802.735336730" observedRunningTime="2025-10-05 08:46:09.759055616 +0000 UTC m=+6803.641682096" watchObservedRunningTime="2025-10-05 08:46:09.763529105 +0000 UTC m=+6803.646155575" Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.289620 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.290169 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.290241 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.291084 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e4a39ab2bd42fbc2828d6cf3524285a042b99f3b26230cc9f73dd09b7781035"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.291197 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://5e4a39ab2bd42fbc2828d6cf3524285a042b99f3b26230cc9f73dd09b7781035" gracePeriod=600 Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.805677 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="5e4a39ab2bd42fbc2828d6cf3524285a042b99f3b26230cc9f73dd09b7781035" exitCode=0 Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.806063 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"5e4a39ab2bd42fbc2828d6cf3524285a042b99f3b26230cc9f73dd09b7781035"} Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.806259 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899"} Oct 05 08:46:14 crc kubenswrapper[4935]: I1005 08:46:14.806287 4935 scope.go:117] "RemoveContainer" containerID="b91e9a598c4de224dd92980a7d777d014e929bb951e0daf54023e5cc2345869b" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.638135 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-84bls"] Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.640749 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.708980 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-84bls"] Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.775196 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-utilities\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.775286 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d22ns\" (UniqueName: \"kubernetes.io/projected/fc7102ef-af12-4949-880f-69c3f13b0aa5-kube-api-access-d22ns\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.775355 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-catalog-content\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.876925 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-catalog-content\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.876986 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-utilities\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.877050 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d22ns\" (UniqueName: \"kubernetes.io/projected/fc7102ef-af12-4949-880f-69c3f13b0aa5-kube-api-access-d22ns\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.877531 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-utilities\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.877531 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-catalog-content\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.903592 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d22ns\" (UniqueName: \"kubernetes.io/projected/fc7102ef-af12-4949-880f-69c3f13b0aa5-kube-api-access-d22ns\") pod \"redhat-marketplace-84bls\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:30 crc kubenswrapper[4935]: I1005 08:47:30.961274 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:31 crc kubenswrapper[4935]: I1005 08:47:31.410751 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-84bls"] Oct 05 08:47:31 crc kubenswrapper[4935]: I1005 08:47:31.527048 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84bls" event={"ID":"fc7102ef-af12-4949-880f-69c3f13b0aa5","Type":"ContainerStarted","Data":"703f397107a42794e004e536171a17b9aad10974f43e0ba0e9cf66a8795198c1"} Oct 05 08:47:32 crc kubenswrapper[4935]: I1005 08:47:32.536781 4935 generic.go:334] "Generic (PLEG): container finished" podID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerID="cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4" exitCode=0 Oct 05 08:47:32 crc kubenswrapper[4935]: I1005 08:47:32.536831 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84bls" event={"ID":"fc7102ef-af12-4949-880f-69c3f13b0aa5","Type":"ContainerDied","Data":"cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4"} Oct 05 08:47:33 crc kubenswrapper[4935]: I1005 08:47:33.548184 4935 generic.go:334] "Generic (PLEG): container finished" podID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerID="ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce" exitCode=0 Oct 05 08:47:33 crc kubenswrapper[4935]: I1005 08:47:33.548486 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84bls" event={"ID":"fc7102ef-af12-4949-880f-69c3f13b0aa5","Type":"ContainerDied","Data":"ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce"} Oct 05 08:47:34 crc kubenswrapper[4935]: I1005 08:47:34.558559 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84bls" event={"ID":"fc7102ef-af12-4949-880f-69c3f13b0aa5","Type":"ContainerStarted","Data":"e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8"} Oct 05 08:47:34 crc kubenswrapper[4935]: I1005 08:47:34.588771 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-84bls" podStartSLOduration=3.048507553 podStartE2EDuration="4.588747861s" podCreationTimestamp="2025-10-05 08:47:30 +0000 UTC" firstStartedPulling="2025-10-05 08:47:32.540065929 +0000 UTC m=+6886.422692389" lastFinishedPulling="2025-10-05 08:47:34.080306227 +0000 UTC m=+6887.962932697" observedRunningTime="2025-10-05 08:47:34.582676999 +0000 UTC m=+6888.465303479" watchObservedRunningTime="2025-10-05 08:47:34.588747861 +0000 UTC m=+6888.471374321" Oct 05 08:47:40 crc kubenswrapper[4935]: I1005 08:47:40.961682 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:40 crc kubenswrapper[4935]: I1005 08:47:40.962593 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.029386 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.560094 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-677g4"] Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.565867 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-677g4" Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.568330 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-677g4"] Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.611745 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6s9\" (UniqueName: \"kubernetes.io/projected/6a2c18d3-0539-408d-acd7-227d032f1787-kube-api-access-9n6s9\") pod \"barbican-db-create-677g4\" (UID: \"6a2c18d3-0539-408d-acd7-227d032f1787\") " pod="openstack/barbican-db-create-677g4" Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.667340 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.712795 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6s9\" (UniqueName: \"kubernetes.io/projected/6a2c18d3-0539-408d-acd7-227d032f1787-kube-api-access-9n6s9\") pod \"barbican-db-create-677g4\" (UID: \"6a2c18d3-0539-408d-acd7-227d032f1787\") " pod="openstack/barbican-db-create-677g4" Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.718091 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-84bls"] Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.733468 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6s9\" (UniqueName: \"kubernetes.io/projected/6a2c18d3-0539-408d-acd7-227d032f1787-kube-api-access-9n6s9\") pod \"barbican-db-create-677g4\" (UID: \"6a2c18d3-0539-408d-acd7-227d032f1787\") " pod="openstack/barbican-db-create-677g4" Oct 05 08:47:41 crc kubenswrapper[4935]: I1005 08:47:41.893808 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-677g4" Oct 05 08:47:42 crc kubenswrapper[4935]: I1005 08:47:42.354165 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-677g4"] Oct 05 08:47:42 crc kubenswrapper[4935]: I1005 08:47:42.637538 4935 generic.go:334] "Generic (PLEG): container finished" podID="6a2c18d3-0539-408d-acd7-227d032f1787" containerID="8779c26fab3cd3f7b5820571169be88c752fd145ae4409ebd8b6fb049b5375e1" exitCode=0 Oct 05 08:47:42 crc kubenswrapper[4935]: I1005 08:47:42.637678 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-677g4" event={"ID":"6a2c18d3-0539-408d-acd7-227d032f1787","Type":"ContainerDied","Data":"8779c26fab3cd3f7b5820571169be88c752fd145ae4409ebd8b6fb049b5375e1"} Oct 05 08:47:42 crc kubenswrapper[4935]: I1005 08:47:42.637946 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-677g4" event={"ID":"6a2c18d3-0539-408d-acd7-227d032f1787","Type":"ContainerStarted","Data":"b5b1c86d0696fac19e1538a8c3fccf05d0ed35700a7ecefa50fbad7362cbf439"} Oct 05 08:47:43 crc kubenswrapper[4935]: I1005 08:47:43.649389 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-84bls" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerName="registry-server" containerID="cri-o://e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8" gracePeriod=2 Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.028340 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-677g4" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.110681 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.156808 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n6s9\" (UniqueName: \"kubernetes.io/projected/6a2c18d3-0539-408d-acd7-227d032f1787-kube-api-access-9n6s9\") pod \"6a2c18d3-0539-408d-acd7-227d032f1787\" (UID: \"6a2c18d3-0539-408d-acd7-227d032f1787\") " Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.163024 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a2c18d3-0539-408d-acd7-227d032f1787-kube-api-access-9n6s9" (OuterVolumeSpecName: "kube-api-access-9n6s9") pod "6a2c18d3-0539-408d-acd7-227d032f1787" (UID: "6a2c18d3-0539-408d-acd7-227d032f1787"). InnerVolumeSpecName "kube-api-access-9n6s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.258836 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d22ns\" (UniqueName: \"kubernetes.io/projected/fc7102ef-af12-4949-880f-69c3f13b0aa5-kube-api-access-d22ns\") pod \"fc7102ef-af12-4949-880f-69c3f13b0aa5\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.259082 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-catalog-content\") pod \"fc7102ef-af12-4949-880f-69c3f13b0aa5\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.259129 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-utilities\") pod \"fc7102ef-af12-4949-880f-69c3f13b0aa5\" (UID: \"fc7102ef-af12-4949-880f-69c3f13b0aa5\") " Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.259412 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n6s9\" (UniqueName: \"kubernetes.io/projected/6a2c18d3-0539-408d-acd7-227d032f1787-kube-api-access-9n6s9\") on node \"crc\" DevicePath \"\"" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.260272 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-utilities" (OuterVolumeSpecName: "utilities") pod "fc7102ef-af12-4949-880f-69c3f13b0aa5" (UID: "fc7102ef-af12-4949-880f-69c3f13b0aa5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.262302 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc7102ef-af12-4949-880f-69c3f13b0aa5-kube-api-access-d22ns" (OuterVolumeSpecName: "kube-api-access-d22ns") pod "fc7102ef-af12-4949-880f-69c3f13b0aa5" (UID: "fc7102ef-af12-4949-880f-69c3f13b0aa5"). InnerVolumeSpecName "kube-api-access-d22ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.276039 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc7102ef-af12-4949-880f-69c3f13b0aa5" (UID: "fc7102ef-af12-4949-880f-69c3f13b0aa5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.361061 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d22ns\" (UniqueName: \"kubernetes.io/projected/fc7102ef-af12-4949-880f-69c3f13b0aa5-kube-api-access-d22ns\") on node \"crc\" DevicePath \"\"" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.361114 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.361133 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc7102ef-af12-4949-880f-69c3f13b0aa5-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.666887 4935 generic.go:334] "Generic (PLEG): container finished" podID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerID="e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8" exitCode=0 Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.667021 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84bls" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.667044 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84bls" event={"ID":"fc7102ef-af12-4949-880f-69c3f13b0aa5","Type":"ContainerDied","Data":"e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8"} Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.667123 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84bls" event={"ID":"fc7102ef-af12-4949-880f-69c3f13b0aa5","Type":"ContainerDied","Data":"703f397107a42794e004e536171a17b9aad10974f43e0ba0e9cf66a8795198c1"} Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.667159 4935 scope.go:117] "RemoveContainer" containerID="e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.670528 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-677g4" event={"ID":"6a2c18d3-0539-408d-acd7-227d032f1787","Type":"ContainerDied","Data":"b5b1c86d0696fac19e1538a8c3fccf05d0ed35700a7ecefa50fbad7362cbf439"} Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.670576 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5b1c86d0696fac19e1538a8c3fccf05d0ed35700a7ecefa50fbad7362cbf439" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.670676 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-677g4" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.728027 4935 scope.go:117] "RemoveContainer" containerID="ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.729362 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-84bls"] Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.740515 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-84bls"] Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.774861 4935 scope.go:117] "RemoveContainer" containerID="cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.791246 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" path="/var/lib/kubelet/pods/fc7102ef-af12-4949-880f-69c3f13b0aa5/volumes" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.817237 4935 scope.go:117] "RemoveContainer" containerID="e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8" Oct 05 08:47:44 crc kubenswrapper[4935]: E1005 08:47:44.817749 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8\": container with ID starting with e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8 not found: ID does not exist" containerID="e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.817812 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8"} err="failed to get container status \"e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8\": rpc error: code = NotFound desc = could not find container \"e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8\": container with ID starting with e21e56cf6441c75dabf53e9648c497a2d592a3739db01873be9ba528875f2cc8 not found: ID does not exist" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.817850 4935 scope.go:117] "RemoveContainer" containerID="ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce" Oct 05 08:47:44 crc kubenswrapper[4935]: E1005 08:47:44.819428 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce\": container with ID starting with ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce not found: ID does not exist" containerID="ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.819491 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce"} err="failed to get container status \"ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce\": rpc error: code = NotFound desc = could not find container \"ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce\": container with ID starting with ff1fdb09e385fc8cb42f6f7d0b0af363479b8cfbb0a719b0aac8278f3c6f91ce not found: ID does not exist" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.819529 4935 scope.go:117] "RemoveContainer" containerID="cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4" Oct 05 08:47:44 crc kubenswrapper[4935]: E1005 08:47:44.820008 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4\": container with ID starting with cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4 not found: ID does not exist" containerID="cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4" Oct 05 08:47:44 crc kubenswrapper[4935]: I1005 08:47:44.820049 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4"} err="failed to get container status \"cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4\": rpc error: code = NotFound desc = could not find container \"cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4\": container with ID starting with cd888fe6c42e0ecafa52b6afbcc10d1e913d0b47cb8cc1b0bb677be2198771a4 not found: ID does not exist" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.669921 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7798-account-create-xhjk2"] Oct 05 08:47:51 crc kubenswrapper[4935]: E1005 08:47:51.670597 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2c18d3-0539-408d-acd7-227d032f1787" containerName="mariadb-database-create" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.670609 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2c18d3-0539-408d-acd7-227d032f1787" containerName="mariadb-database-create" Oct 05 08:47:51 crc kubenswrapper[4935]: E1005 08:47:51.670619 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerName="extract-utilities" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.670625 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerName="extract-utilities" Oct 05 08:47:51 crc kubenswrapper[4935]: E1005 08:47:51.670635 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerName="extract-content" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.670642 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerName="extract-content" Oct 05 08:47:51 crc kubenswrapper[4935]: E1005 08:47:51.670660 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerName="registry-server" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.670666 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerName="registry-server" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.670811 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc7102ef-af12-4949-880f-69c3f13b0aa5" containerName="registry-server" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.670832 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2c18d3-0539-408d-acd7-227d032f1787" containerName="mariadb-database-create" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.671351 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7798-account-create-xhjk2" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.674629 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.700842 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7798-account-create-xhjk2"] Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.780587 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh582\" (UniqueName: \"kubernetes.io/projected/00f0f894-2eef-4b8a-9e38-3e71e7c0283a-kube-api-access-nh582\") pod \"barbican-7798-account-create-xhjk2\" (UID: \"00f0f894-2eef-4b8a-9e38-3e71e7c0283a\") " pod="openstack/barbican-7798-account-create-xhjk2" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.884016 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh582\" (UniqueName: \"kubernetes.io/projected/00f0f894-2eef-4b8a-9e38-3e71e7c0283a-kube-api-access-nh582\") pod \"barbican-7798-account-create-xhjk2\" (UID: \"00f0f894-2eef-4b8a-9e38-3e71e7c0283a\") " pod="openstack/barbican-7798-account-create-xhjk2" Oct 05 08:47:51 crc kubenswrapper[4935]: I1005 08:47:51.904539 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh582\" (UniqueName: \"kubernetes.io/projected/00f0f894-2eef-4b8a-9e38-3e71e7c0283a-kube-api-access-nh582\") pod \"barbican-7798-account-create-xhjk2\" (UID: \"00f0f894-2eef-4b8a-9e38-3e71e7c0283a\") " pod="openstack/barbican-7798-account-create-xhjk2" Oct 05 08:47:52 crc kubenswrapper[4935]: I1005 08:47:52.060074 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7798-account-create-xhjk2" Oct 05 08:47:52 crc kubenswrapper[4935]: I1005 08:47:52.315170 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7798-account-create-xhjk2"] Oct 05 08:47:52 crc kubenswrapper[4935]: I1005 08:47:52.772550 4935 generic.go:334] "Generic (PLEG): container finished" podID="00f0f894-2eef-4b8a-9e38-3e71e7c0283a" containerID="d454166d0eb85993d4ae1fd650b5c910c14f648dca66d693b01e551a36d34764" exitCode=0 Oct 05 08:47:52 crc kubenswrapper[4935]: I1005 08:47:52.772676 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7798-account-create-xhjk2" event={"ID":"00f0f894-2eef-4b8a-9e38-3e71e7c0283a","Type":"ContainerDied","Data":"d454166d0eb85993d4ae1fd650b5c910c14f648dca66d693b01e551a36d34764"} Oct 05 08:47:52 crc kubenswrapper[4935]: I1005 08:47:52.772832 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7798-account-create-xhjk2" event={"ID":"00f0f894-2eef-4b8a-9e38-3e71e7c0283a","Type":"ContainerStarted","Data":"396af855835c45afd08ffa0310e0ee2ed6cbd8262f6d74267cd912bf90ae842c"} Oct 05 08:47:54 crc kubenswrapper[4935]: I1005 08:47:54.151869 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7798-account-create-xhjk2" Oct 05 08:47:54 crc kubenswrapper[4935]: I1005 08:47:54.329978 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh582\" (UniqueName: \"kubernetes.io/projected/00f0f894-2eef-4b8a-9e38-3e71e7c0283a-kube-api-access-nh582\") pod \"00f0f894-2eef-4b8a-9e38-3e71e7c0283a\" (UID: \"00f0f894-2eef-4b8a-9e38-3e71e7c0283a\") " Oct 05 08:47:54 crc kubenswrapper[4935]: I1005 08:47:54.336244 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f0f894-2eef-4b8a-9e38-3e71e7c0283a-kube-api-access-nh582" (OuterVolumeSpecName: "kube-api-access-nh582") pod "00f0f894-2eef-4b8a-9e38-3e71e7c0283a" (UID: "00f0f894-2eef-4b8a-9e38-3e71e7c0283a"). InnerVolumeSpecName "kube-api-access-nh582". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:47:54 crc kubenswrapper[4935]: I1005 08:47:54.432725 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh582\" (UniqueName: \"kubernetes.io/projected/00f0f894-2eef-4b8a-9e38-3e71e7c0283a-kube-api-access-nh582\") on node \"crc\" DevicePath \"\"" Oct 05 08:47:54 crc kubenswrapper[4935]: I1005 08:47:54.797836 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7798-account-create-xhjk2" event={"ID":"00f0f894-2eef-4b8a-9e38-3e71e7c0283a","Type":"ContainerDied","Data":"396af855835c45afd08ffa0310e0ee2ed6cbd8262f6d74267cd912bf90ae842c"} Oct 05 08:47:54 crc kubenswrapper[4935]: I1005 08:47:54.797977 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="396af855835c45afd08ffa0310e0ee2ed6cbd8262f6d74267cd912bf90ae842c" Oct 05 08:47:54 crc kubenswrapper[4935]: I1005 08:47:54.797980 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7798-account-create-xhjk2" Oct 05 08:47:56 crc kubenswrapper[4935]: I1005 08:47:56.901703 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-r2sf9"] Oct 05 08:47:56 crc kubenswrapper[4935]: E1005 08:47:56.902576 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f0f894-2eef-4b8a-9e38-3e71e7c0283a" containerName="mariadb-account-create" Oct 05 08:47:56 crc kubenswrapper[4935]: I1005 08:47:56.902600 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f0f894-2eef-4b8a-9e38-3e71e7c0283a" containerName="mariadb-account-create" Oct 05 08:47:56 crc kubenswrapper[4935]: I1005 08:47:56.902933 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f0f894-2eef-4b8a-9e38-3e71e7c0283a" containerName="mariadb-account-create" Oct 05 08:47:56 crc kubenswrapper[4935]: I1005 08:47:56.903795 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:56 crc kubenswrapper[4935]: I1005 08:47:56.906635 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-7sgjs" Oct 05 08:47:56 crc kubenswrapper[4935]: I1005 08:47:56.907006 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 08:47:56 crc kubenswrapper[4935]: I1005 08:47:56.911642 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-r2sf9"] Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.078430 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-combined-ca-bundle\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.078851 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-db-sync-config-data\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.078927 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpzhl\" (UniqueName: \"kubernetes.io/projected/7e2eb559-8cea-4191-9f10-5b564959ac9b-kube-api-access-tpzhl\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.180118 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpzhl\" (UniqueName: \"kubernetes.io/projected/7e2eb559-8cea-4191-9f10-5b564959ac9b-kube-api-access-tpzhl\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.180273 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-combined-ca-bundle\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.180358 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-db-sync-config-data\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.188415 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-combined-ca-bundle\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.189936 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-db-sync-config-data\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.202157 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpzhl\" (UniqueName: \"kubernetes.io/projected/7e2eb559-8cea-4191-9f10-5b564959ac9b-kube-api-access-tpzhl\") pod \"barbican-db-sync-r2sf9\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.224151 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.751524 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-r2sf9"] Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.754679 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:47:57 crc kubenswrapper[4935]: I1005 08:47:57.829819 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r2sf9" event={"ID":"7e2eb559-8cea-4191-9f10-5b564959ac9b","Type":"ContainerStarted","Data":"b75f1a13f669e2dcd63a213a131c32305d214977725348cd5d0c3f8975b02689"} Oct 05 08:48:03 crc kubenswrapper[4935]: I1005 08:48:03.897025 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r2sf9" event={"ID":"7e2eb559-8cea-4191-9f10-5b564959ac9b","Type":"ContainerStarted","Data":"afc6a31d527b8f997602ffb199a324e35da96ba522ed08d42332ed13903bd713"} Oct 05 08:48:03 crc kubenswrapper[4935]: I1005 08:48:03.933010 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-r2sf9" podStartSLOduration=2.752136672 podStartE2EDuration="7.932983827s" podCreationTimestamp="2025-10-05 08:47:56 +0000 UTC" firstStartedPulling="2025-10-05 08:47:57.754412834 +0000 UTC m=+6911.637039294" lastFinishedPulling="2025-10-05 08:48:02.935259959 +0000 UTC m=+6916.817886449" observedRunningTime="2025-10-05 08:48:03.919155859 +0000 UTC m=+6917.801782389" watchObservedRunningTime="2025-10-05 08:48:03.932983827 +0000 UTC m=+6917.815610327" Oct 05 08:48:05 crc kubenswrapper[4935]: I1005 08:48:05.922397 4935 generic.go:334] "Generic (PLEG): container finished" podID="7e2eb559-8cea-4191-9f10-5b564959ac9b" containerID="afc6a31d527b8f997602ffb199a324e35da96ba522ed08d42332ed13903bd713" exitCode=0 Oct 05 08:48:05 crc kubenswrapper[4935]: I1005 08:48:05.922499 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r2sf9" event={"ID":"7e2eb559-8cea-4191-9f10-5b564959ac9b","Type":"ContainerDied","Data":"afc6a31d527b8f997602ffb199a324e35da96ba522ed08d42332ed13903bd713"} Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.256442 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.378449 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-db-sync-config-data\") pod \"7e2eb559-8cea-4191-9f10-5b564959ac9b\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.378678 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpzhl\" (UniqueName: \"kubernetes.io/projected/7e2eb559-8cea-4191-9f10-5b564959ac9b-kube-api-access-tpzhl\") pod \"7e2eb559-8cea-4191-9f10-5b564959ac9b\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.378750 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-combined-ca-bundle\") pod \"7e2eb559-8cea-4191-9f10-5b564959ac9b\" (UID: \"7e2eb559-8cea-4191-9f10-5b564959ac9b\") " Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.383981 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e2eb559-8cea-4191-9f10-5b564959ac9b-kube-api-access-tpzhl" (OuterVolumeSpecName: "kube-api-access-tpzhl") pod "7e2eb559-8cea-4191-9f10-5b564959ac9b" (UID: "7e2eb559-8cea-4191-9f10-5b564959ac9b"). InnerVolumeSpecName "kube-api-access-tpzhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.385204 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7e2eb559-8cea-4191-9f10-5b564959ac9b" (UID: "7e2eb559-8cea-4191-9f10-5b564959ac9b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.426318 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e2eb559-8cea-4191-9f10-5b564959ac9b" (UID: "7e2eb559-8cea-4191-9f10-5b564959ac9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.480474 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpzhl\" (UniqueName: \"kubernetes.io/projected/7e2eb559-8cea-4191-9f10-5b564959ac9b-kube-api-access-tpzhl\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.480514 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.480526 4935 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e2eb559-8cea-4191-9f10-5b564959ac9b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.966513 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r2sf9" event={"ID":"7e2eb559-8cea-4191-9f10-5b564959ac9b","Type":"ContainerDied","Data":"b75f1a13f669e2dcd63a213a131c32305d214977725348cd5d0c3f8975b02689"} Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.966558 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b75f1a13f669e2dcd63a213a131c32305d214977725348cd5d0c3f8975b02689" Oct 05 08:48:07 crc kubenswrapper[4935]: I1005 08:48:07.966655 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r2sf9" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.211579 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5c558b76b7-7vtbc"] Oct 05 08:48:08 crc kubenswrapper[4935]: E1005 08:48:08.212024 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2eb559-8cea-4191-9f10-5b564959ac9b" containerName="barbican-db-sync" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.212055 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2eb559-8cea-4191-9f10-5b564959ac9b" containerName="barbican-db-sync" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.212273 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e2eb559-8cea-4191-9f10-5b564959ac9b" containerName="barbican-db-sync" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.214023 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.217730 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.218105 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.218276 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-7sgjs" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.230130 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c558b76b7-7vtbc"] Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.266592 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-57cf597c6d-gng7h"] Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.268125 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.276680 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.294747 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-logs\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.294789 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mwtd\" (UniqueName: \"kubernetes.io/projected/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-kube-api-access-5mwtd\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.294877 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-config-data\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.294916 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-config-data-custom\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.294936 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-combined-ca-bundle\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.323622 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-57cf597c6d-gng7h"] Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.362628 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6994fc6f49-dn579"] Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.383796 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.398795 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-combined-ca-bundle\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.398974 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-logs\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399007 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mwtd\" (UniqueName: \"kubernetes.io/projected/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-kube-api-access-5mwtd\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399047 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-config-data-custom\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399091 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-config-data\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399115 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jct5\" (UniqueName: \"kubernetes.io/projected/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-kube-api-access-4jct5\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399207 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-config-data\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399242 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-config-data-custom\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399268 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-combined-ca-bundle\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399291 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-logs\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.399361 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-logs\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.400867 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6994fc6f49-dn579"] Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.404047 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-config-data-custom\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.433135 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-config-data\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.434858 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-combined-ca-bundle\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.466607 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mwtd\" (UniqueName: \"kubernetes.io/projected/dbb4faf3-8787-4ca3-a956-6794a9ad86ea-kube-api-access-5mwtd\") pod \"barbican-worker-5c558b76b7-7vtbc\" (UID: \"dbb4faf3-8787-4ca3-a956-6794a9ad86ea\") " pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.498463 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-85d576456b-gxmw2"] Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500469 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-config\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500515 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-dns-svc\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500549 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-logs\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500597 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-combined-ca-bundle\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500678 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-nb\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500735 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-config-data-custom\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500782 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-config-data\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500799 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jct5\" (UniqueName: \"kubernetes.io/projected/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-kube-api-access-4jct5\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.500821 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-sb\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.501061 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g959\" (UniqueName: \"kubernetes.io/projected/6efae153-e456-4e1b-a04f-5caf84761718-kube-api-access-7g959\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.501272 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-logs\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.502013 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.505255 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.517419 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85d576456b-gxmw2"] Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.525666 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-config-data-custom\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.533690 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-combined-ca-bundle\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.538159 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-config-data\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.550339 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c558b76b7-7vtbc" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.557625 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jct5\" (UniqueName: \"kubernetes.io/projected/e4c4f39e-ff2a-48a2-acb5-df5542a6e218-kube-api-access-4jct5\") pod \"barbican-keystone-listener-57cf597c6d-gng7h\" (UID: \"e4c4f39e-ff2a-48a2-acb5-df5542a6e218\") " pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605326 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-config-data-custom\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605371 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-config\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605394 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-dns-svc\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605416 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mppvq\" (UniqueName: \"kubernetes.io/projected/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-kube-api-access-mppvq\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605455 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-combined-ca-bundle\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605488 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-nb\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605509 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-logs\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605528 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-config-data\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605563 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-sb\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.605586 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g959\" (UniqueName: \"kubernetes.io/projected/6efae153-e456-4e1b-a04f-5caf84761718-kube-api-access-7g959\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.609164 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-config\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.609874 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-dns-svc\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.610056 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-sb\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.615332 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-nb\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.631509 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.652870 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g959\" (UniqueName: \"kubernetes.io/projected/6efae153-e456-4e1b-a04f-5caf84761718-kube-api-access-7g959\") pod \"dnsmasq-dns-6994fc6f49-dn579\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.706909 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-combined-ca-bundle\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.706981 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-logs\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.707006 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-config-data\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.707107 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-config-data-custom\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.707141 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mppvq\" (UniqueName: \"kubernetes.io/projected/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-kube-api-access-mppvq\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.711416 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-logs\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.719015 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-combined-ca-bundle\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.719052 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-config-data-custom\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.719749 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-config-data\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.730942 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mppvq\" (UniqueName: \"kubernetes.io/projected/dd1c0eed-cccf-453d-8e01-59ce9890f7eb-kube-api-access-mppvq\") pod \"barbican-api-85d576456b-gxmw2\" (UID: \"dd1c0eed-cccf-453d-8e01-59ce9890f7eb\") " pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.790969 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:08 crc kubenswrapper[4935]: I1005 08:48:08.959321 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:09 crc kubenswrapper[4935]: I1005 08:48:09.118993 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c558b76b7-7vtbc"] Oct 05 08:48:09 crc kubenswrapper[4935]: W1005 08:48:09.125005 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbb4faf3_8787_4ca3_a956_6794a9ad86ea.slice/crio-2be52fbe59b2dedfc023f4468a470ce05e81247ea6c7af0c3d9766e79439c620 WatchSource:0}: Error finding container 2be52fbe59b2dedfc023f4468a470ce05e81247ea6c7af0c3d9766e79439c620: Status 404 returned error can't find the container with id 2be52fbe59b2dedfc023f4468a470ce05e81247ea6c7af0c3d9766e79439c620 Oct 05 08:48:09 crc kubenswrapper[4935]: I1005 08:48:09.171805 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-57cf597c6d-gng7h"] Oct 05 08:48:09 crc kubenswrapper[4935]: I1005 08:48:09.273372 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6994fc6f49-dn579"] Oct 05 08:48:09 crc kubenswrapper[4935]: W1005 08:48:09.278004 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6efae153_e456_4e1b_a04f_5caf84761718.slice/crio-db3d621e831a2232937ae6deb142c248c0c18d211c635d291c1e89c09dddef57 WatchSource:0}: Error finding container db3d621e831a2232937ae6deb142c248c0c18d211c635d291c1e89c09dddef57: Status 404 returned error can't find the container with id db3d621e831a2232937ae6deb142c248c0c18d211c635d291c1e89c09dddef57 Oct 05 08:48:09 crc kubenswrapper[4935]: I1005 08:48:09.441005 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85d576456b-gxmw2"] Oct 05 08:48:09 crc kubenswrapper[4935]: I1005 08:48:09.997786 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c558b76b7-7vtbc" event={"ID":"dbb4faf3-8787-4ca3-a956-6794a9ad86ea","Type":"ContainerStarted","Data":"2be52fbe59b2dedfc023f4468a470ce05e81247ea6c7af0c3d9766e79439c620"} Oct 05 08:48:09 crc kubenswrapper[4935]: I1005 08:48:09.999025 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" event={"ID":"e4c4f39e-ff2a-48a2-acb5-df5542a6e218","Type":"ContainerStarted","Data":"5485302a44e3626c292697d1426fd82102fc2745473085baf27dbd0a0eb9c4bc"} Oct 05 08:48:10 crc kubenswrapper[4935]: I1005 08:48:10.000317 4935 generic.go:334] "Generic (PLEG): container finished" podID="6efae153-e456-4e1b-a04f-5caf84761718" containerID="21f5649b65aa66bc0cdb34ebf3628b79e4a6d6fd33b08888e7cbab3ce5b67c40" exitCode=0 Oct 05 08:48:10 crc kubenswrapper[4935]: I1005 08:48:10.000357 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" event={"ID":"6efae153-e456-4e1b-a04f-5caf84761718","Type":"ContainerDied","Data":"21f5649b65aa66bc0cdb34ebf3628b79e4a6d6fd33b08888e7cbab3ce5b67c40"} Oct 05 08:48:10 crc kubenswrapper[4935]: I1005 08:48:10.000373 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" event={"ID":"6efae153-e456-4e1b-a04f-5caf84761718","Type":"ContainerStarted","Data":"db3d621e831a2232937ae6deb142c248c0c18d211c635d291c1e89c09dddef57"} Oct 05 08:48:10 crc kubenswrapper[4935]: I1005 08:48:10.002938 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d576456b-gxmw2" event={"ID":"dd1c0eed-cccf-453d-8e01-59ce9890f7eb","Type":"ContainerStarted","Data":"0679d790fb31daa8b8259f584bdac22702d9f1f67119a3bc14df37a1e81a8862"} Oct 05 08:48:10 crc kubenswrapper[4935]: I1005 08:48:10.002968 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d576456b-gxmw2" event={"ID":"dd1c0eed-cccf-453d-8e01-59ce9890f7eb","Type":"ContainerStarted","Data":"b2a1495ef08a560dabb6beccba29037220f2172697abc13302322a4efe586e0d"} Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.012799 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d576456b-gxmw2" event={"ID":"dd1c0eed-cccf-453d-8e01-59ce9890f7eb","Type":"ContainerStarted","Data":"08ffbfee2adff7e5a2c211a01fcfd4deeed5976d6246e2076630b5526d8d8e87"} Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.013098 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.013110 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.014746 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c558b76b7-7vtbc" event={"ID":"dbb4faf3-8787-4ca3-a956-6794a9ad86ea","Type":"ContainerStarted","Data":"7a6236720359ad5beb02fbe99f4ac7648f4677ce4a1111d67c7acec2066524ce"} Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.014793 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c558b76b7-7vtbc" event={"ID":"dbb4faf3-8787-4ca3-a956-6794a9ad86ea","Type":"ContainerStarted","Data":"9b56629f9194b619c8170b4c625a8ebf6d479ed9ccaecfda820e28e412b815f2"} Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.016665 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" event={"ID":"e4c4f39e-ff2a-48a2-acb5-df5542a6e218","Type":"ContainerStarted","Data":"f5ac1b02388e9c3d1223e7366d8d13cf053fc221613adb2e2cb0583da4bd2ef3"} Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.016740 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" event={"ID":"e4c4f39e-ff2a-48a2-acb5-df5542a6e218","Type":"ContainerStarted","Data":"718f06c7580112b28654eaffb86818779a861d3d994668b55cab3cc17ee52d09"} Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.018028 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" event={"ID":"6efae153-e456-4e1b-a04f-5caf84761718","Type":"ContainerStarted","Data":"c7055ef8b0d480c8b737e73faeb9be4df7153079b1e0fcb0fe9d395f44d108d1"} Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.018746 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.054642 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-57cf597c6d-gng7h" podStartSLOduration=1.856189233 podStartE2EDuration="3.054626924s" podCreationTimestamp="2025-10-05 08:48:08 +0000 UTC" firstStartedPulling="2025-10-05 08:48:09.181555965 +0000 UTC m=+6923.064182425" lastFinishedPulling="2025-10-05 08:48:10.379993646 +0000 UTC m=+6924.262620116" observedRunningTime="2025-10-05 08:48:11.05335979 +0000 UTC m=+6924.935986260" watchObservedRunningTime="2025-10-05 08:48:11.054626924 +0000 UTC m=+6924.937253384" Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.056559 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-85d576456b-gxmw2" podStartSLOduration=3.056553085 podStartE2EDuration="3.056553085s" podCreationTimestamp="2025-10-05 08:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:48:11.034508298 +0000 UTC m=+6924.917134758" watchObservedRunningTime="2025-10-05 08:48:11.056553085 +0000 UTC m=+6924.939179545" Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.075321 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5c558b76b7-7vtbc" podStartSLOduration=1.8215387299999999 podStartE2EDuration="3.075302024s" podCreationTimestamp="2025-10-05 08:48:08 +0000 UTC" firstStartedPulling="2025-10-05 08:48:09.126685204 +0000 UTC m=+6923.009311664" lastFinishedPulling="2025-10-05 08:48:10.380448498 +0000 UTC m=+6924.263074958" observedRunningTime="2025-10-05 08:48:11.071667757 +0000 UTC m=+6924.954294217" watchObservedRunningTime="2025-10-05 08:48:11.075302024 +0000 UTC m=+6924.957928484" Oct 05 08:48:11 crc kubenswrapper[4935]: I1005 08:48:11.088838 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" podStartSLOduration=3.088818514 podStartE2EDuration="3.088818514s" podCreationTimestamp="2025-10-05 08:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:48:11.087709434 +0000 UTC m=+6924.970335894" watchObservedRunningTime="2025-10-05 08:48:11.088818514 +0000 UTC m=+6924.971444984" Oct 05 08:48:14 crc kubenswrapper[4935]: I1005 08:48:14.289756 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:48:14 crc kubenswrapper[4935]: I1005 08:48:14.290272 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:48:18 crc kubenswrapper[4935]: I1005 08:48:18.793100 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:48:18 crc kubenswrapper[4935]: I1005 08:48:18.881472 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c8c8bc55-xbt6n"] Oct 05 08:48:18 crc kubenswrapper[4935]: I1005 08:48:18.881740 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" podUID="aa95a826-eeee-4268-a765-a755c5bc349c" containerName="dnsmasq-dns" containerID="cri-o://1d57c2318d21c9fcf6ce9760e9eaeee91bb3b8ee734fecfc5623918793bce431" gracePeriod=10 Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.119610 4935 generic.go:334] "Generic (PLEG): container finished" podID="aa95a826-eeee-4268-a765-a755c5bc349c" containerID="1d57c2318d21c9fcf6ce9760e9eaeee91bb3b8ee734fecfc5623918793bce431" exitCode=0 Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.119848 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" event={"ID":"aa95a826-eeee-4268-a765-a755c5bc349c","Type":"ContainerDied","Data":"1d57c2318d21c9fcf6ce9760e9eaeee91bb3b8ee734fecfc5623918793bce431"} Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.441510 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.550660 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5dsh\" (UniqueName: \"kubernetes.io/projected/aa95a826-eeee-4268-a765-a755c5bc349c-kube-api-access-p5dsh\") pod \"aa95a826-eeee-4268-a765-a755c5bc349c\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.551006 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-sb\") pod \"aa95a826-eeee-4268-a765-a755c5bc349c\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.551059 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-config\") pod \"aa95a826-eeee-4268-a765-a755c5bc349c\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.551096 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-dns-svc\") pod \"aa95a826-eeee-4268-a765-a755c5bc349c\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.551195 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-nb\") pod \"aa95a826-eeee-4268-a765-a755c5bc349c\" (UID: \"aa95a826-eeee-4268-a765-a755c5bc349c\") " Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.576854 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa95a826-eeee-4268-a765-a755c5bc349c-kube-api-access-p5dsh" (OuterVolumeSpecName: "kube-api-access-p5dsh") pod "aa95a826-eeee-4268-a765-a755c5bc349c" (UID: "aa95a826-eeee-4268-a765-a755c5bc349c"). InnerVolumeSpecName "kube-api-access-p5dsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.594596 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-config" (OuterVolumeSpecName: "config") pod "aa95a826-eeee-4268-a765-a755c5bc349c" (UID: "aa95a826-eeee-4268-a765-a755c5bc349c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.606132 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa95a826-eeee-4268-a765-a755c5bc349c" (UID: "aa95a826-eeee-4268-a765-a755c5bc349c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.607749 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa95a826-eeee-4268-a765-a755c5bc349c" (UID: "aa95a826-eeee-4268-a765-a755c5bc349c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.619401 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa95a826-eeee-4268-a765-a755c5bc349c" (UID: "aa95a826-eeee-4268-a765-a755c5bc349c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.658092 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5dsh\" (UniqueName: \"kubernetes.io/projected/aa95a826-eeee-4268-a765-a755c5bc349c-kube-api-access-p5dsh\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.658472 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.659253 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.659351 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:19 crc kubenswrapper[4935]: I1005 08:48:19.659416 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa95a826-eeee-4268-a765-a755c5bc349c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.131253 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" event={"ID":"aa95a826-eeee-4268-a765-a755c5bc349c","Type":"ContainerDied","Data":"f6bfb341599b9ee34ae0ee4b171607edf4f6b5d127772873aa49be03ff0a6ff2"} Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.131666 4935 scope.go:117] "RemoveContainer" containerID="1d57c2318d21c9fcf6ce9760e9eaeee91bb3b8ee734fecfc5623918793bce431" Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.131501 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c8c8bc55-xbt6n" Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.162810 4935 scope.go:117] "RemoveContainer" containerID="99f0c5c556ddcb36370a192dbbcb4368af41583ae8347670d9a10607598d7bc4" Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.183517 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c8c8bc55-xbt6n"] Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.192683 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54c8c8bc55-xbt6n"] Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.374934 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.456715 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85d576456b-gxmw2" Oct 05 08:48:20 crc kubenswrapper[4935]: I1005 08:48:20.787220 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa95a826-eeee-4268-a765-a755c5bc349c" path="/var/lib/kubelet/pods/aa95a826-eeee-4268-a765-a755c5bc349c/volumes" Oct 05 08:48:27 crc kubenswrapper[4935]: I1005 08:48:27.963076 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-fmvtl"] Oct 05 08:48:27 crc kubenswrapper[4935]: E1005 08:48:27.964285 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa95a826-eeee-4268-a765-a755c5bc349c" containerName="dnsmasq-dns" Oct 05 08:48:27 crc kubenswrapper[4935]: I1005 08:48:27.964309 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa95a826-eeee-4268-a765-a755c5bc349c" containerName="dnsmasq-dns" Oct 05 08:48:27 crc kubenswrapper[4935]: E1005 08:48:27.964340 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa95a826-eeee-4268-a765-a755c5bc349c" containerName="init" Oct 05 08:48:27 crc kubenswrapper[4935]: I1005 08:48:27.964356 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa95a826-eeee-4268-a765-a755c5bc349c" containerName="init" Oct 05 08:48:27 crc kubenswrapper[4935]: I1005 08:48:27.964674 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa95a826-eeee-4268-a765-a755c5bc349c" containerName="dnsmasq-dns" Oct 05 08:48:27 crc kubenswrapper[4935]: I1005 08:48:27.965627 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fmvtl" Oct 05 08:48:27 crc kubenswrapper[4935]: I1005 08:48:27.972003 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fmvtl"] Oct 05 08:48:28 crc kubenswrapper[4935]: I1005 08:48:28.118171 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxk84\" (UniqueName: \"kubernetes.io/projected/77f7d891-b3c6-4649-9596-f4b3132b47e0-kube-api-access-jxk84\") pod \"neutron-db-create-fmvtl\" (UID: \"77f7d891-b3c6-4649-9596-f4b3132b47e0\") " pod="openstack/neutron-db-create-fmvtl" Oct 05 08:48:28 crc kubenswrapper[4935]: I1005 08:48:28.225344 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxk84\" (UniqueName: \"kubernetes.io/projected/77f7d891-b3c6-4649-9596-f4b3132b47e0-kube-api-access-jxk84\") pod \"neutron-db-create-fmvtl\" (UID: \"77f7d891-b3c6-4649-9596-f4b3132b47e0\") " pod="openstack/neutron-db-create-fmvtl" Oct 05 08:48:28 crc kubenswrapper[4935]: I1005 08:48:28.265158 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxk84\" (UniqueName: \"kubernetes.io/projected/77f7d891-b3c6-4649-9596-f4b3132b47e0-kube-api-access-jxk84\") pod \"neutron-db-create-fmvtl\" (UID: \"77f7d891-b3c6-4649-9596-f4b3132b47e0\") " pod="openstack/neutron-db-create-fmvtl" Oct 05 08:48:28 crc kubenswrapper[4935]: I1005 08:48:28.287504 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fmvtl" Oct 05 08:48:28 crc kubenswrapper[4935]: I1005 08:48:28.774783 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fmvtl"] Oct 05 08:48:29 crc kubenswrapper[4935]: I1005 08:48:29.266241 4935 generic.go:334] "Generic (PLEG): container finished" podID="77f7d891-b3c6-4649-9596-f4b3132b47e0" containerID="135e09dd83ca1be0f4c18e6764f986b4108307c6c20807a70682b635819d1bcd" exitCode=0 Oct 05 08:48:29 crc kubenswrapper[4935]: I1005 08:48:29.266327 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fmvtl" event={"ID":"77f7d891-b3c6-4649-9596-f4b3132b47e0","Type":"ContainerDied","Data":"135e09dd83ca1be0f4c18e6764f986b4108307c6c20807a70682b635819d1bcd"} Oct 05 08:48:29 crc kubenswrapper[4935]: I1005 08:48:29.266386 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fmvtl" event={"ID":"77f7d891-b3c6-4649-9596-f4b3132b47e0","Type":"ContainerStarted","Data":"93565e4fdeb2e3c3812a2c390b442f24ff4f006c6270cafb37cfff1bc3283e62"} Oct 05 08:48:31 crc kubenswrapper[4935]: I1005 08:48:30.606181 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fmvtl" Oct 05 08:48:31 crc kubenswrapper[4935]: I1005 08:48:30.772748 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxk84\" (UniqueName: \"kubernetes.io/projected/77f7d891-b3c6-4649-9596-f4b3132b47e0-kube-api-access-jxk84\") pod \"77f7d891-b3c6-4649-9596-f4b3132b47e0\" (UID: \"77f7d891-b3c6-4649-9596-f4b3132b47e0\") " Oct 05 08:48:31 crc kubenswrapper[4935]: I1005 08:48:30.778766 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f7d891-b3c6-4649-9596-f4b3132b47e0-kube-api-access-jxk84" (OuterVolumeSpecName: "kube-api-access-jxk84") pod "77f7d891-b3c6-4649-9596-f4b3132b47e0" (UID: "77f7d891-b3c6-4649-9596-f4b3132b47e0"). InnerVolumeSpecName "kube-api-access-jxk84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:48:31 crc kubenswrapper[4935]: I1005 08:48:30.874457 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxk84\" (UniqueName: \"kubernetes.io/projected/77f7d891-b3c6-4649-9596-f4b3132b47e0-kube-api-access-jxk84\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:31 crc kubenswrapper[4935]: I1005 08:48:31.297486 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fmvtl" event={"ID":"77f7d891-b3c6-4649-9596-f4b3132b47e0","Type":"ContainerDied","Data":"93565e4fdeb2e3c3812a2c390b442f24ff4f006c6270cafb37cfff1bc3283e62"} Oct 05 08:48:31 crc kubenswrapper[4935]: I1005 08:48:31.300945 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93565e4fdeb2e3c3812a2c390b442f24ff4f006c6270cafb37cfff1bc3283e62" Oct 05 08:48:31 crc kubenswrapper[4935]: I1005 08:48:31.297566 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fmvtl" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.112989 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d00d-account-create-vc9qx"] Oct 05 08:48:38 crc kubenswrapper[4935]: E1005 08:48:38.113970 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f7d891-b3c6-4649-9596-f4b3132b47e0" containerName="mariadb-database-create" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.113991 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f7d891-b3c6-4649-9596-f4b3132b47e0" containerName="mariadb-database-create" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.114212 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f7d891-b3c6-4649-9596-f4b3132b47e0" containerName="mariadb-database-create" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.114925 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d00d-account-create-vc9qx" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.117828 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.122704 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d00d-account-create-vc9qx"] Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.297079 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52mlc\" (UniqueName: \"kubernetes.io/projected/ec8da432-4737-4b96-97bb-7e163a809f75-kube-api-access-52mlc\") pod \"neutron-d00d-account-create-vc9qx\" (UID: \"ec8da432-4737-4b96-97bb-7e163a809f75\") " pod="openstack/neutron-d00d-account-create-vc9qx" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.399315 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52mlc\" (UniqueName: \"kubernetes.io/projected/ec8da432-4737-4b96-97bb-7e163a809f75-kube-api-access-52mlc\") pod \"neutron-d00d-account-create-vc9qx\" (UID: \"ec8da432-4737-4b96-97bb-7e163a809f75\") " pod="openstack/neutron-d00d-account-create-vc9qx" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.436127 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52mlc\" (UniqueName: \"kubernetes.io/projected/ec8da432-4737-4b96-97bb-7e163a809f75-kube-api-access-52mlc\") pod \"neutron-d00d-account-create-vc9qx\" (UID: \"ec8da432-4737-4b96-97bb-7e163a809f75\") " pod="openstack/neutron-d00d-account-create-vc9qx" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.442277 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d00d-account-create-vc9qx" Oct 05 08:48:38 crc kubenswrapper[4935]: I1005 08:48:38.870267 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d00d-account-create-vc9qx"] Oct 05 08:48:38 crc kubenswrapper[4935]: W1005 08:48:38.878748 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec8da432_4737_4b96_97bb_7e163a809f75.slice/crio-bf2ca4f11a5f14b5cd44eea7949404facfd87c2d12b3567854b80500ea254f8f WatchSource:0}: Error finding container bf2ca4f11a5f14b5cd44eea7949404facfd87c2d12b3567854b80500ea254f8f: Status 404 returned error can't find the container with id bf2ca4f11a5f14b5cd44eea7949404facfd87c2d12b3567854b80500ea254f8f Oct 05 08:48:39 crc kubenswrapper[4935]: I1005 08:48:39.381679 4935 generic.go:334] "Generic (PLEG): container finished" podID="ec8da432-4737-4b96-97bb-7e163a809f75" containerID="cce76dcc3fa400c3a3b0819021f5ac69981210a476f595caa7c2dc2fe7e616b5" exitCode=0 Oct 05 08:48:39 crc kubenswrapper[4935]: I1005 08:48:39.381745 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d00d-account-create-vc9qx" event={"ID":"ec8da432-4737-4b96-97bb-7e163a809f75","Type":"ContainerDied","Data":"cce76dcc3fa400c3a3b0819021f5ac69981210a476f595caa7c2dc2fe7e616b5"} Oct 05 08:48:39 crc kubenswrapper[4935]: I1005 08:48:39.381777 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d00d-account-create-vc9qx" event={"ID":"ec8da432-4737-4b96-97bb-7e163a809f75","Type":"ContainerStarted","Data":"bf2ca4f11a5f14b5cd44eea7949404facfd87c2d12b3567854b80500ea254f8f"} Oct 05 08:48:40 crc kubenswrapper[4935]: I1005 08:48:40.763372 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d00d-account-create-vc9qx" Oct 05 08:48:40 crc kubenswrapper[4935]: I1005 08:48:40.941612 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52mlc\" (UniqueName: \"kubernetes.io/projected/ec8da432-4737-4b96-97bb-7e163a809f75-kube-api-access-52mlc\") pod \"ec8da432-4737-4b96-97bb-7e163a809f75\" (UID: \"ec8da432-4737-4b96-97bb-7e163a809f75\") " Oct 05 08:48:40 crc kubenswrapper[4935]: I1005 08:48:40.948390 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec8da432-4737-4b96-97bb-7e163a809f75-kube-api-access-52mlc" (OuterVolumeSpecName: "kube-api-access-52mlc") pod "ec8da432-4737-4b96-97bb-7e163a809f75" (UID: "ec8da432-4737-4b96-97bb-7e163a809f75"). InnerVolumeSpecName "kube-api-access-52mlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:48:41 crc kubenswrapper[4935]: I1005 08:48:41.043649 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52mlc\" (UniqueName: \"kubernetes.io/projected/ec8da432-4737-4b96-97bb-7e163a809f75-kube-api-access-52mlc\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:41 crc kubenswrapper[4935]: I1005 08:48:41.401522 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d00d-account-create-vc9qx" event={"ID":"ec8da432-4737-4b96-97bb-7e163a809f75","Type":"ContainerDied","Data":"bf2ca4f11a5f14b5cd44eea7949404facfd87c2d12b3567854b80500ea254f8f"} Oct 05 08:48:41 crc kubenswrapper[4935]: I1005 08:48:41.401565 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf2ca4f11a5f14b5cd44eea7949404facfd87c2d12b3567854b80500ea254f8f" Oct 05 08:48:41 crc kubenswrapper[4935]: I1005 08:48:41.401625 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d00d-account-create-vc9qx" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.295478 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-d7twt"] Oct 05 08:48:43 crc kubenswrapper[4935]: E1005 08:48:43.296303 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8da432-4737-4b96-97bb-7e163a809f75" containerName="mariadb-account-create" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.296326 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8da432-4737-4b96-97bb-7e163a809f75" containerName="mariadb-account-create" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.296561 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8da432-4737-4b96-97bb-7e163a809f75" containerName="mariadb-account-create" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.297455 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.300571 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.300707 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.302055 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r7gg5" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.312855 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-d7twt"] Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.487650 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-config\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.487751 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-combined-ca-bundle\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.487779 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk2vx\" (UniqueName: \"kubernetes.io/projected/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-kube-api-access-nk2vx\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.590574 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-combined-ca-bundle\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.590659 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk2vx\" (UniqueName: \"kubernetes.io/projected/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-kube-api-access-nk2vx\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.590823 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-config\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.598645 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-combined-ca-bundle\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.598732 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-config\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.613324 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk2vx\" (UniqueName: \"kubernetes.io/projected/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-kube-api-access-nk2vx\") pod \"neutron-db-sync-d7twt\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:43 crc kubenswrapper[4935]: I1005 08:48:43.634355 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:44 crc kubenswrapper[4935]: I1005 08:48:44.250654 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-d7twt"] Oct 05 08:48:44 crc kubenswrapper[4935]: I1005 08:48:44.289807 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:48:44 crc kubenswrapper[4935]: I1005 08:48:44.289873 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:48:44 crc kubenswrapper[4935]: I1005 08:48:44.454720 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d7twt" event={"ID":"a03a073a-7848-4ba1-b73a-ed3afa62a1e8","Type":"ContainerStarted","Data":"5eadd1cd946c5586d843a75abad297459c95b85f169b14714f69c2dc39e19b72"} Oct 05 08:48:44 crc kubenswrapper[4935]: I1005 08:48:44.454781 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d7twt" event={"ID":"a03a073a-7848-4ba1-b73a-ed3afa62a1e8","Type":"ContainerStarted","Data":"72a827bccc94379bd457567e682822d517f08405e2788b7c1995bd455e7d1142"} Oct 05 08:48:44 crc kubenswrapper[4935]: I1005 08:48:44.469051 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-d7twt" podStartSLOduration=1.469033027 podStartE2EDuration="1.469033027s" podCreationTimestamp="2025-10-05 08:48:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:48:44.466717986 +0000 UTC m=+6958.349344456" watchObservedRunningTime="2025-10-05 08:48:44.469033027 +0000 UTC m=+6958.351659487" Oct 05 08:48:48 crc kubenswrapper[4935]: I1005 08:48:48.510957 4935 generic.go:334] "Generic (PLEG): container finished" podID="a03a073a-7848-4ba1-b73a-ed3afa62a1e8" containerID="5eadd1cd946c5586d843a75abad297459c95b85f169b14714f69c2dc39e19b72" exitCode=0 Oct 05 08:48:48 crc kubenswrapper[4935]: I1005 08:48:48.511054 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d7twt" event={"ID":"a03a073a-7848-4ba1-b73a-ed3afa62a1e8","Type":"ContainerDied","Data":"5eadd1cd946c5586d843a75abad297459c95b85f169b14714f69c2dc39e19b72"} Oct 05 08:48:49 crc kubenswrapper[4935]: I1005 08:48:49.856022 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.023682 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk2vx\" (UniqueName: \"kubernetes.io/projected/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-kube-api-access-nk2vx\") pod \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.023881 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-combined-ca-bundle\") pod \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.023942 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-config\") pod \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\" (UID: \"a03a073a-7848-4ba1-b73a-ed3afa62a1e8\") " Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.029387 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-kube-api-access-nk2vx" (OuterVolumeSpecName: "kube-api-access-nk2vx") pod "a03a073a-7848-4ba1-b73a-ed3afa62a1e8" (UID: "a03a073a-7848-4ba1-b73a-ed3afa62a1e8"). InnerVolumeSpecName "kube-api-access-nk2vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.046741 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-config" (OuterVolumeSpecName: "config") pod "a03a073a-7848-4ba1-b73a-ed3afa62a1e8" (UID: "a03a073a-7848-4ba1-b73a-ed3afa62a1e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.058473 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a03a073a-7848-4ba1-b73a-ed3afa62a1e8" (UID: "a03a073a-7848-4ba1-b73a-ed3afa62a1e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.125286 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.125320 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.125330 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk2vx\" (UniqueName: \"kubernetes.io/projected/a03a073a-7848-4ba1-b73a-ed3afa62a1e8-kube-api-access-nk2vx\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.543923 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d7twt" event={"ID":"a03a073a-7848-4ba1-b73a-ed3afa62a1e8","Type":"ContainerDied","Data":"72a827bccc94379bd457567e682822d517f08405e2788b7c1995bd455e7d1142"} Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.543989 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72a827bccc94379bd457567e682822d517f08405e2788b7c1995bd455e7d1142" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.544151 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d7twt" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.838175 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8576ff4fd9-wl24b"] Oct 05 08:48:50 crc kubenswrapper[4935]: E1005 08:48:50.838850 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03a073a-7848-4ba1-b73a-ed3afa62a1e8" containerName="neutron-db-sync" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.838874 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03a073a-7848-4ba1-b73a-ed3afa62a1e8" containerName="neutron-db-sync" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.839084 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a03a073a-7848-4ba1-b73a-ed3afa62a1e8" containerName="neutron-db-sync" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.840083 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.859621 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8576ff4fd9-wl24b"] Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.872350 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-79b64455c-9bh2f"] Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.877778 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.882289 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.882346 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r7gg5" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.883082 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.905508 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79b64455c-9bh2f"] Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.940809 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-nb\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.941045 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-config\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.941168 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt2lc\" (UniqueName: \"kubernetes.io/projected/b3d3be9b-6d3c-4174-96f9-7e975624879b-kube-api-access-jt2lc\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.941310 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-dns-svc\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:50 crc kubenswrapper[4935]: I1005 08:48:50.941371 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-sb\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.043340 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt2lc\" (UniqueName: \"kubernetes.io/projected/b3d3be9b-6d3c-4174-96f9-7e975624879b-kube-api-access-jt2lc\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.043527 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-dns-svc\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.043556 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-sb\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.044702 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-sb\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.044780 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-dns-svc\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.044873 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-combined-ca-bundle\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.045804 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tftlx\" (UniqueName: \"kubernetes.io/projected/6c373000-c6d0-4a7d-9623-2804446b1507-kube-api-access-tftlx\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.045942 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-nb\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.046064 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-config\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.046128 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-httpd-config\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.046225 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-config\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.046563 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-nb\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.046807 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-config\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.064925 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt2lc\" (UniqueName: \"kubernetes.io/projected/b3d3be9b-6d3c-4174-96f9-7e975624879b-kube-api-access-jt2lc\") pod \"dnsmasq-dns-8576ff4fd9-wl24b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.148073 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-combined-ca-bundle\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.148136 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tftlx\" (UniqueName: \"kubernetes.io/projected/6c373000-c6d0-4a7d-9623-2804446b1507-kube-api-access-tftlx\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.148181 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-httpd-config\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.148209 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-config\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.157736 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-httpd-config\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.157917 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-combined-ca-bundle\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.158476 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6c373000-c6d0-4a7d-9623-2804446b1507-config\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.167187 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.168165 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tftlx\" (UniqueName: \"kubernetes.io/projected/6c373000-c6d0-4a7d-9623-2804446b1507-kube-api-access-tftlx\") pod \"neutron-79b64455c-9bh2f\" (UID: \"6c373000-c6d0-4a7d-9623-2804446b1507\") " pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.201236 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.649677 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8576ff4fd9-wl24b"] Oct 05 08:48:51 crc kubenswrapper[4935]: I1005 08:48:51.827227 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79b64455c-9bh2f"] Oct 05 08:48:51 crc kubenswrapper[4935]: W1005 08:48:51.827295 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c373000_c6d0_4a7d_9623_2804446b1507.slice/crio-056dd447325fe9b55f54973966c2fecd9c874e6d937d4965652774e78f5170fd WatchSource:0}: Error finding container 056dd447325fe9b55f54973966c2fecd9c874e6d937d4965652774e78f5170fd: Status 404 returned error can't find the container with id 056dd447325fe9b55f54973966c2fecd9c874e6d937d4965652774e78f5170fd Oct 05 08:48:52 crc kubenswrapper[4935]: I1005 08:48:52.563421 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b64455c-9bh2f" event={"ID":"6c373000-c6d0-4a7d-9623-2804446b1507","Type":"ContainerStarted","Data":"3ba7c49cac7a09bad233f735d398f0052483abc214427e3a903b61745ea9b1a5"} Oct 05 08:48:52 crc kubenswrapper[4935]: I1005 08:48:52.563822 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:48:52 crc kubenswrapper[4935]: I1005 08:48:52.563840 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b64455c-9bh2f" event={"ID":"6c373000-c6d0-4a7d-9623-2804446b1507","Type":"ContainerStarted","Data":"0bd7bb51ffdaae433d951460646a63568413380341b2f2e3239e5f0bd8394bbc"} Oct 05 08:48:52 crc kubenswrapper[4935]: I1005 08:48:52.563855 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b64455c-9bh2f" event={"ID":"6c373000-c6d0-4a7d-9623-2804446b1507","Type":"ContainerStarted","Data":"056dd447325fe9b55f54973966c2fecd9c874e6d937d4965652774e78f5170fd"} Oct 05 08:48:52 crc kubenswrapper[4935]: I1005 08:48:52.564907 4935 generic.go:334] "Generic (PLEG): container finished" podID="b3d3be9b-6d3c-4174-96f9-7e975624879b" containerID="b37606a2cedf01304a3c640f175efdff0be220d9a7e53715a94f96416f94100d" exitCode=0 Oct 05 08:48:52 crc kubenswrapper[4935]: I1005 08:48:52.564955 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" event={"ID":"b3d3be9b-6d3c-4174-96f9-7e975624879b","Type":"ContainerDied","Data":"b37606a2cedf01304a3c640f175efdff0be220d9a7e53715a94f96416f94100d"} Oct 05 08:48:52 crc kubenswrapper[4935]: I1005 08:48:52.564998 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" event={"ID":"b3d3be9b-6d3c-4174-96f9-7e975624879b","Type":"ContainerStarted","Data":"70f90adec5ae6376cd8bbfb6eeca6cc03dfa9a14f290a1dad80359b729fdbd3e"} Oct 05 08:48:52 crc kubenswrapper[4935]: I1005 08:48:52.591178 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-79b64455c-9bh2f" podStartSLOduration=2.591152024 podStartE2EDuration="2.591152024s" podCreationTimestamp="2025-10-05 08:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:48:52.58423003 +0000 UTC m=+6966.466856520" watchObservedRunningTime="2025-10-05 08:48:52.591152024 +0000 UTC m=+6966.473778484" Oct 05 08:48:53 crc kubenswrapper[4935]: I1005 08:48:53.590328 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" event={"ID":"b3d3be9b-6d3c-4174-96f9-7e975624879b","Type":"ContainerStarted","Data":"af49540a5f86e4963231c61955cc5c74b72559c01cc2534a0e06874ca33fe137"} Oct 05 08:48:53 crc kubenswrapper[4935]: I1005 08:48:53.590712 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:48:53 crc kubenswrapper[4935]: I1005 08:48:53.627067 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" podStartSLOduration=3.6270374480000003 podStartE2EDuration="3.627037448s" podCreationTimestamp="2025-10-05 08:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:48:53.61358581 +0000 UTC m=+6967.496212360" watchObservedRunningTime="2025-10-05 08:48:53.627037448 +0000 UTC m=+6967.509663928" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.170216 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.288422 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6994fc6f49-dn579"] Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.288698 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" podUID="6efae153-e456-4e1b-a04f-5caf84761718" containerName="dnsmasq-dns" containerID="cri-o://c7055ef8b0d480c8b737e73faeb9be4df7153079b1e0fcb0fe9d395f44d108d1" gracePeriod=10 Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.680794 4935 generic.go:334] "Generic (PLEG): container finished" podID="6efae153-e456-4e1b-a04f-5caf84761718" containerID="c7055ef8b0d480c8b737e73faeb9be4df7153079b1e0fcb0fe9d395f44d108d1" exitCode=0 Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.681195 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" event={"ID":"6efae153-e456-4e1b-a04f-5caf84761718","Type":"ContainerDied","Data":"c7055ef8b0d480c8b737e73faeb9be4df7153079b1e0fcb0fe9d395f44d108d1"} Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.681225 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" event={"ID":"6efae153-e456-4e1b-a04f-5caf84761718","Type":"ContainerDied","Data":"db3d621e831a2232937ae6deb142c248c0c18d211c635d291c1e89c09dddef57"} Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.681237 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db3d621e831a2232937ae6deb142c248c0c18d211c635d291c1e89c09dddef57" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.731452 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.859068 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-config\") pod \"6efae153-e456-4e1b-a04f-5caf84761718\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.859596 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-dns-svc\") pod \"6efae153-e456-4e1b-a04f-5caf84761718\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.859626 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-sb\") pod \"6efae153-e456-4e1b-a04f-5caf84761718\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.859659 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g959\" (UniqueName: \"kubernetes.io/projected/6efae153-e456-4e1b-a04f-5caf84761718-kube-api-access-7g959\") pod \"6efae153-e456-4e1b-a04f-5caf84761718\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.859682 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-nb\") pod \"6efae153-e456-4e1b-a04f-5caf84761718\" (UID: \"6efae153-e456-4e1b-a04f-5caf84761718\") " Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.869734 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6efae153-e456-4e1b-a04f-5caf84761718-kube-api-access-7g959" (OuterVolumeSpecName: "kube-api-access-7g959") pod "6efae153-e456-4e1b-a04f-5caf84761718" (UID: "6efae153-e456-4e1b-a04f-5caf84761718"). InnerVolumeSpecName "kube-api-access-7g959". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.898733 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6efae153-e456-4e1b-a04f-5caf84761718" (UID: "6efae153-e456-4e1b-a04f-5caf84761718"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.906457 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6efae153-e456-4e1b-a04f-5caf84761718" (UID: "6efae153-e456-4e1b-a04f-5caf84761718"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.906460 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6efae153-e456-4e1b-a04f-5caf84761718" (UID: "6efae153-e456-4e1b-a04f-5caf84761718"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.907497 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-config" (OuterVolumeSpecName: "config") pod "6efae153-e456-4e1b-a04f-5caf84761718" (UID: "6efae153-e456-4e1b-a04f-5caf84761718"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.961184 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.961217 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.961239 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.961255 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g959\" (UniqueName: \"kubernetes.io/projected/6efae153-e456-4e1b-a04f-5caf84761718-kube-api-access-7g959\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:01 crc kubenswrapper[4935]: I1005 08:49:01.961268 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6efae153-e456-4e1b-a04f-5caf84761718-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:02 crc kubenswrapper[4935]: I1005 08:49:02.690139 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6994fc6f49-dn579" Oct 05 08:49:02 crc kubenswrapper[4935]: I1005 08:49:02.734714 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6994fc6f49-dn579"] Oct 05 08:49:02 crc kubenswrapper[4935]: I1005 08:49:02.744056 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6994fc6f49-dn579"] Oct 05 08:49:02 crc kubenswrapper[4935]: I1005 08:49:02.787440 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6efae153-e456-4e1b-a04f-5caf84761718" path="/var/lib/kubelet/pods/6efae153-e456-4e1b-a04f-5caf84761718/volumes" Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.289360 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.290065 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.290141 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.290869 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.290953 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" gracePeriod=600 Oct 05 08:49:14 crc kubenswrapper[4935]: E1005 08:49:14.410278 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.810388 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" exitCode=0 Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.810440 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899"} Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.810474 4935 scope.go:117] "RemoveContainer" containerID="5e4a39ab2bd42fbc2828d6cf3524285a042b99f3b26230cc9f73dd09b7781035" Oct 05 08:49:14 crc kubenswrapper[4935]: I1005 08:49:14.811032 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:49:14 crc kubenswrapper[4935]: E1005 08:49:14.811337 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:49:21 crc kubenswrapper[4935]: I1005 08:49:21.223081 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-79b64455c-9bh2f" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.714346 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-f849r"] Oct 05 08:49:28 crc kubenswrapper[4935]: E1005 08:49:28.715405 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6efae153-e456-4e1b-a04f-5caf84761718" containerName="init" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.715420 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6efae153-e456-4e1b-a04f-5caf84761718" containerName="init" Oct 05 08:49:28 crc kubenswrapper[4935]: E1005 08:49:28.715449 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6efae153-e456-4e1b-a04f-5caf84761718" containerName="dnsmasq-dns" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.715455 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6efae153-e456-4e1b-a04f-5caf84761718" containerName="dnsmasq-dns" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.715635 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6efae153-e456-4e1b-a04f-5caf84761718" containerName="dnsmasq-dns" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.716247 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f849r" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.723913 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-f849r"] Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.755934 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kzmm\" (UniqueName: \"kubernetes.io/projected/b87e8ba5-cc47-41b7-b1dd-155a2942ce6a-kube-api-access-2kzmm\") pod \"glance-db-create-f849r\" (UID: \"b87e8ba5-cc47-41b7-b1dd-155a2942ce6a\") " pod="openstack/glance-db-create-f849r" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.778910 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:49:28 crc kubenswrapper[4935]: E1005 08:49:28.779268 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.858311 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kzmm\" (UniqueName: \"kubernetes.io/projected/b87e8ba5-cc47-41b7-b1dd-155a2942ce6a-kube-api-access-2kzmm\") pod \"glance-db-create-f849r\" (UID: \"b87e8ba5-cc47-41b7-b1dd-155a2942ce6a\") " pod="openstack/glance-db-create-f849r" Oct 05 08:49:28 crc kubenswrapper[4935]: I1005 08:49:28.875962 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kzmm\" (UniqueName: \"kubernetes.io/projected/b87e8ba5-cc47-41b7-b1dd-155a2942ce6a-kube-api-access-2kzmm\") pod \"glance-db-create-f849r\" (UID: \"b87e8ba5-cc47-41b7-b1dd-155a2942ce6a\") " pod="openstack/glance-db-create-f849r" Oct 05 08:49:29 crc kubenswrapper[4935]: I1005 08:49:29.036097 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f849r" Oct 05 08:49:29 crc kubenswrapper[4935]: I1005 08:49:29.575792 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-f849r"] Oct 05 08:49:29 crc kubenswrapper[4935]: W1005 08:49:29.585141 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb87e8ba5_cc47_41b7_b1dd_155a2942ce6a.slice/crio-2ac91c10d278db593d3fa267ac5adddf2212552656722a1d9aeda3287f1debf1 WatchSource:0}: Error finding container 2ac91c10d278db593d3fa267ac5adddf2212552656722a1d9aeda3287f1debf1: Status 404 returned error can't find the container with id 2ac91c10d278db593d3fa267ac5adddf2212552656722a1d9aeda3287f1debf1 Oct 05 08:49:29 crc kubenswrapper[4935]: I1005 08:49:29.962942 4935 generic.go:334] "Generic (PLEG): container finished" podID="b87e8ba5-cc47-41b7-b1dd-155a2942ce6a" containerID="7a5141487b7219c38654c0be6915513a17cd06b9d7b917b24d2fdb28949e18e2" exitCode=0 Oct 05 08:49:29 crc kubenswrapper[4935]: I1005 08:49:29.963001 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f849r" event={"ID":"b87e8ba5-cc47-41b7-b1dd-155a2942ce6a","Type":"ContainerDied","Data":"7a5141487b7219c38654c0be6915513a17cd06b9d7b917b24d2fdb28949e18e2"} Oct 05 08:49:29 crc kubenswrapper[4935]: I1005 08:49:29.963041 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f849r" event={"ID":"b87e8ba5-cc47-41b7-b1dd-155a2942ce6a","Type":"ContainerStarted","Data":"2ac91c10d278db593d3fa267ac5adddf2212552656722a1d9aeda3287f1debf1"} Oct 05 08:49:31 crc kubenswrapper[4935]: I1005 08:49:31.386230 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f849r" Oct 05 08:49:31 crc kubenswrapper[4935]: I1005 08:49:31.406159 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kzmm\" (UniqueName: \"kubernetes.io/projected/b87e8ba5-cc47-41b7-b1dd-155a2942ce6a-kube-api-access-2kzmm\") pod \"b87e8ba5-cc47-41b7-b1dd-155a2942ce6a\" (UID: \"b87e8ba5-cc47-41b7-b1dd-155a2942ce6a\") " Oct 05 08:49:31 crc kubenswrapper[4935]: I1005 08:49:31.414084 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b87e8ba5-cc47-41b7-b1dd-155a2942ce6a-kube-api-access-2kzmm" (OuterVolumeSpecName: "kube-api-access-2kzmm") pod "b87e8ba5-cc47-41b7-b1dd-155a2942ce6a" (UID: "b87e8ba5-cc47-41b7-b1dd-155a2942ce6a"). InnerVolumeSpecName "kube-api-access-2kzmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:49:31 crc kubenswrapper[4935]: I1005 08:49:31.508277 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kzmm\" (UniqueName: \"kubernetes.io/projected/b87e8ba5-cc47-41b7-b1dd-155a2942ce6a-kube-api-access-2kzmm\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:31 crc kubenswrapper[4935]: I1005 08:49:31.990548 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f849r" event={"ID":"b87e8ba5-cc47-41b7-b1dd-155a2942ce6a","Type":"ContainerDied","Data":"2ac91c10d278db593d3fa267ac5adddf2212552656722a1d9aeda3287f1debf1"} Oct 05 08:49:31 crc kubenswrapper[4935]: I1005 08:49:31.990611 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ac91c10d278db593d3fa267ac5adddf2212552656722a1d9aeda3287f1debf1" Oct 05 08:49:31 crc kubenswrapper[4935]: I1005 08:49:31.990612 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f849r" Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.793063 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-3d00-account-create-hdc7z"] Oct 05 08:49:38 crc kubenswrapper[4935]: E1005 08:49:38.793828 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b87e8ba5-cc47-41b7-b1dd-155a2942ce6a" containerName="mariadb-database-create" Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.793841 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b87e8ba5-cc47-41b7-b1dd-155a2942ce6a" containerName="mariadb-database-create" Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.794027 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b87e8ba5-cc47-41b7-b1dd-155a2942ce6a" containerName="mariadb-database-create" Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.794530 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3d00-account-create-hdc7z" Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.796317 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.803619 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3d00-account-create-hdc7z"] Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.837058 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfcr8\" (UniqueName: \"kubernetes.io/projected/ae0958d8-c9b8-4c10-b422-a66922f5c96b-kube-api-access-gfcr8\") pod \"glance-3d00-account-create-hdc7z\" (UID: \"ae0958d8-c9b8-4c10-b422-a66922f5c96b\") " pod="openstack/glance-3d00-account-create-hdc7z" Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.938260 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfcr8\" (UniqueName: \"kubernetes.io/projected/ae0958d8-c9b8-4c10-b422-a66922f5c96b-kube-api-access-gfcr8\") pod \"glance-3d00-account-create-hdc7z\" (UID: \"ae0958d8-c9b8-4c10-b422-a66922f5c96b\") " pod="openstack/glance-3d00-account-create-hdc7z" Oct 05 08:49:38 crc kubenswrapper[4935]: I1005 08:49:38.958067 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfcr8\" (UniqueName: \"kubernetes.io/projected/ae0958d8-c9b8-4c10-b422-a66922f5c96b-kube-api-access-gfcr8\") pod \"glance-3d00-account-create-hdc7z\" (UID: \"ae0958d8-c9b8-4c10-b422-a66922f5c96b\") " pod="openstack/glance-3d00-account-create-hdc7z" Oct 05 08:49:39 crc kubenswrapper[4935]: I1005 08:49:39.115797 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3d00-account-create-hdc7z" Oct 05 08:49:39 crc kubenswrapper[4935]: I1005 08:49:39.711104 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3d00-account-create-hdc7z"] Oct 05 08:49:39 crc kubenswrapper[4935]: I1005 08:49:39.777601 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:49:39 crc kubenswrapper[4935]: E1005 08:49:39.778045 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:49:40 crc kubenswrapper[4935]: I1005 08:49:40.066707 4935 generic.go:334] "Generic (PLEG): container finished" podID="ae0958d8-c9b8-4c10-b422-a66922f5c96b" containerID="3eeb54cb99164ce98b1ba3a478e7c2bf72a7b055ccaf933aefdc8b4e9da97cff" exitCode=0 Oct 05 08:49:40 crc kubenswrapper[4935]: I1005 08:49:40.067040 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3d00-account-create-hdc7z" event={"ID":"ae0958d8-c9b8-4c10-b422-a66922f5c96b","Type":"ContainerDied","Data":"3eeb54cb99164ce98b1ba3a478e7c2bf72a7b055ccaf933aefdc8b4e9da97cff"} Oct 05 08:49:40 crc kubenswrapper[4935]: I1005 08:49:40.067067 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3d00-account-create-hdc7z" event={"ID":"ae0958d8-c9b8-4c10-b422-a66922f5c96b","Type":"ContainerStarted","Data":"8387f230b778eb890bca6dc606940fa09a5cbc8e1a3950527835aa615eb40391"} Oct 05 08:49:41 crc kubenswrapper[4935]: I1005 08:49:41.497282 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3d00-account-create-hdc7z" Oct 05 08:49:41 crc kubenswrapper[4935]: I1005 08:49:41.686175 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfcr8\" (UniqueName: \"kubernetes.io/projected/ae0958d8-c9b8-4c10-b422-a66922f5c96b-kube-api-access-gfcr8\") pod \"ae0958d8-c9b8-4c10-b422-a66922f5c96b\" (UID: \"ae0958d8-c9b8-4c10-b422-a66922f5c96b\") " Oct 05 08:49:41 crc kubenswrapper[4935]: I1005 08:49:41.691583 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae0958d8-c9b8-4c10-b422-a66922f5c96b-kube-api-access-gfcr8" (OuterVolumeSpecName: "kube-api-access-gfcr8") pod "ae0958d8-c9b8-4c10-b422-a66922f5c96b" (UID: "ae0958d8-c9b8-4c10-b422-a66922f5c96b"). InnerVolumeSpecName "kube-api-access-gfcr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:49:41 crc kubenswrapper[4935]: I1005 08:49:41.788325 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfcr8\" (UniqueName: \"kubernetes.io/projected/ae0958d8-c9b8-4c10-b422-a66922f5c96b-kube-api-access-gfcr8\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:42 crc kubenswrapper[4935]: I1005 08:49:42.090746 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3d00-account-create-hdc7z" event={"ID":"ae0958d8-c9b8-4c10-b422-a66922f5c96b","Type":"ContainerDied","Data":"8387f230b778eb890bca6dc606940fa09a5cbc8e1a3950527835aa615eb40391"} Oct 05 08:49:42 crc kubenswrapper[4935]: I1005 08:49:42.090805 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8387f230b778eb890bca6dc606940fa09a5cbc8e1a3950527835aa615eb40391" Oct 05 08:49:42 crc kubenswrapper[4935]: I1005 08:49:42.090931 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3d00-account-create-hdc7z" Oct 05 08:49:43 crc kubenswrapper[4935]: I1005 08:49:43.955801 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-rqppz"] Oct 05 08:49:43 crc kubenswrapper[4935]: E1005 08:49:43.956590 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae0958d8-c9b8-4c10-b422-a66922f5c96b" containerName="mariadb-account-create" Oct 05 08:49:43 crc kubenswrapper[4935]: I1005 08:49:43.956609 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae0958d8-c9b8-4c10-b422-a66922f5c96b" containerName="mariadb-account-create" Oct 05 08:49:43 crc kubenswrapper[4935]: I1005 08:49:43.956845 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae0958d8-c9b8-4c10-b422-a66922f5c96b" containerName="mariadb-account-create" Oct 05 08:49:43 crc kubenswrapper[4935]: I1005 08:49:43.957856 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:43 crc kubenswrapper[4935]: I1005 08:49:43.966828 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 05 08:49:43 crc kubenswrapper[4935]: I1005 08:49:43.966895 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dz7vr" Oct 05 08:49:43 crc kubenswrapper[4935]: I1005 08:49:43.974162 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rqppz"] Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.128933 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-config-data\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.129045 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-db-sync-config-data\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.129072 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcf82\" (UniqueName: \"kubernetes.io/projected/0657582d-ba22-485a-9964-efb3e4b41fe7-kube-api-access-qcf82\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.129091 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-combined-ca-bundle\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.230581 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-config-data\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.230708 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-db-sync-config-data\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.230738 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcf82\" (UniqueName: \"kubernetes.io/projected/0657582d-ba22-485a-9964-efb3e4b41fe7-kube-api-access-qcf82\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.230762 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-combined-ca-bundle\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.243508 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-db-sync-config-data\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.243665 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-config-data\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.243863 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-combined-ca-bundle\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.271803 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcf82\" (UniqueName: \"kubernetes.io/projected/0657582d-ba22-485a-9964-efb3e4b41fe7-kube-api-access-qcf82\") pod \"glance-db-sync-rqppz\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.285353 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rqppz" Oct 05 08:49:44 crc kubenswrapper[4935]: I1005 08:49:44.875314 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rqppz"] Oct 05 08:49:45 crc kubenswrapper[4935]: I1005 08:49:45.115063 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rqppz" event={"ID":"0657582d-ba22-485a-9964-efb3e4b41fe7","Type":"ContainerStarted","Data":"567e63ffe925fcae57e3a7f59922a9a7858508e9c06a492ddf7e292df1c8ac1b"} Oct 05 08:49:51 crc kubenswrapper[4935]: I1005 08:49:51.777073 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:49:51 crc kubenswrapper[4935]: E1005 08:49:51.777793 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:49:57 crc kubenswrapper[4935]: I1005 08:49:57.379456 4935 scope.go:117] "RemoveContainer" containerID="80a35954bc1fe1b70a90a2ca5f6ae6b64172125e8a9480bdf87409babde50cb1" Oct 05 08:50:00 crc kubenswrapper[4935]: I1005 08:50:00.236977 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rqppz" event={"ID":"0657582d-ba22-485a-9964-efb3e4b41fe7","Type":"ContainerStarted","Data":"a93fa822736fd8b2807064fd05ce98c2aae5ae3d7cf223827499f4add0414a4e"} Oct 05 08:50:00 crc kubenswrapper[4935]: I1005 08:50:00.261444 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-rqppz" podStartSLOduration=3.004978619 podStartE2EDuration="17.261398388s" podCreationTimestamp="2025-10-05 08:49:43 +0000 UTC" firstStartedPulling="2025-10-05 08:49:44.87738651 +0000 UTC m=+7018.760012970" lastFinishedPulling="2025-10-05 08:49:59.133806279 +0000 UTC m=+7033.016432739" observedRunningTime="2025-10-05 08:50:00.255736858 +0000 UTC m=+7034.138363368" watchObservedRunningTime="2025-10-05 08:50:00.261398388 +0000 UTC m=+7034.144024858" Oct 05 08:50:03 crc kubenswrapper[4935]: I1005 08:50:03.269395 4935 generic.go:334] "Generic (PLEG): container finished" podID="0657582d-ba22-485a-9964-efb3e4b41fe7" containerID="a93fa822736fd8b2807064fd05ce98c2aae5ae3d7cf223827499f4add0414a4e" exitCode=0 Oct 05 08:50:03 crc kubenswrapper[4935]: I1005 08:50:03.269571 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rqppz" event={"ID":"0657582d-ba22-485a-9964-efb3e4b41fe7","Type":"ContainerDied","Data":"a93fa822736fd8b2807064fd05ce98c2aae5ae3d7cf223827499f4add0414a4e"} Oct 05 08:50:04 crc kubenswrapper[4935]: I1005 08:50:04.777991 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:50:04 crc kubenswrapper[4935]: E1005 08:50:04.778520 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:50:04 crc kubenswrapper[4935]: I1005 08:50:04.845008 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rqppz" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.001769 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-config-data\") pod \"0657582d-ba22-485a-9964-efb3e4b41fe7\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.002080 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcf82\" (UniqueName: \"kubernetes.io/projected/0657582d-ba22-485a-9964-efb3e4b41fe7-kube-api-access-qcf82\") pod \"0657582d-ba22-485a-9964-efb3e4b41fe7\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.002136 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-combined-ca-bundle\") pod \"0657582d-ba22-485a-9964-efb3e4b41fe7\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.002211 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-db-sync-config-data\") pod \"0657582d-ba22-485a-9964-efb3e4b41fe7\" (UID: \"0657582d-ba22-485a-9964-efb3e4b41fe7\") " Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.007772 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0657582d-ba22-485a-9964-efb3e4b41fe7-kube-api-access-qcf82" (OuterVolumeSpecName: "kube-api-access-qcf82") pod "0657582d-ba22-485a-9964-efb3e4b41fe7" (UID: "0657582d-ba22-485a-9964-efb3e4b41fe7"). InnerVolumeSpecName "kube-api-access-qcf82". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.020033 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0657582d-ba22-485a-9964-efb3e4b41fe7" (UID: "0657582d-ba22-485a-9964-efb3e4b41fe7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.033879 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0657582d-ba22-485a-9964-efb3e4b41fe7" (UID: "0657582d-ba22-485a-9964-efb3e4b41fe7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.073936 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-config-data" (OuterVolumeSpecName: "config-data") pod "0657582d-ba22-485a-9964-efb3e4b41fe7" (UID: "0657582d-ba22-485a-9964-efb3e4b41fe7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.104472 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.104522 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcf82\" (UniqueName: \"kubernetes.io/projected/0657582d-ba22-485a-9964-efb3e4b41fe7-kube-api-access-qcf82\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.104543 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.104561 4935 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0657582d-ba22-485a-9964-efb3e4b41fe7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.291659 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rqppz" event={"ID":"0657582d-ba22-485a-9964-efb3e4b41fe7","Type":"ContainerDied","Data":"567e63ffe925fcae57e3a7f59922a9a7858508e9c06a492ddf7e292df1c8ac1b"} Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.291711 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="567e63ffe925fcae57e3a7f59922a9a7858508e9c06a492ddf7e292df1c8ac1b" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.291762 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rqppz" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.701441 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:05 crc kubenswrapper[4935]: E1005 08:50:05.701877 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0657582d-ba22-485a-9964-efb3e4b41fe7" containerName="glance-db-sync" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.701910 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0657582d-ba22-485a-9964-efb3e4b41fe7" containerName="glance-db-sync" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.702134 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0657582d-ba22-485a-9964-efb3e4b41fe7" containerName="glance-db-sync" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.703311 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.708154 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.708265 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.708396 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.708509 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dz7vr" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.766116 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.820737 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.820798 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-scripts\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.820828 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvgh6\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-kube-api-access-zvgh6\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.820855 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-ceph\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.820913 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-config-data\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.820933 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-logs\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.820989 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.860398 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784c8875b9-kv9rb"] Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.873129 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.884297 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784c8875b9-kv9rb"] Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.923741 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.923840 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.923865 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-scripts\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.923911 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvgh6\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-kube-api-access-zvgh6\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.923938 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-ceph\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.925069 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-config-data\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.925121 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-logs\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.925594 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-logs\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.929404 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-config-data\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.929961 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-ceph\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.935280 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.936364 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.942660 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-scripts\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:05 crc kubenswrapper[4935]: I1005 08:50:05.988260 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvgh6\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-kube-api-access-zvgh6\") pod \"glance-default-external-api-0\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.006065 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.009395 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.013231 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.016371 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026410 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-dns-svc\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026470 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bnsk\" (UniqueName: \"kubernetes.io/projected/f7c9038f-342a-4d92-b283-086b6a0c7e7e-kube-api-access-5bnsk\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026517 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026533 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-logs\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026559 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026595 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026627 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-nb\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026651 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026679 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j58tp\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-kube-api-access-j58tp\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026697 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-sb\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026718 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-config\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.026739 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.035420 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.128954 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129018 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-nb\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129043 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129078 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j58tp\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-kube-api-access-j58tp\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129102 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-sb\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129130 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-config\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129144 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129182 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-dns-svc\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129225 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bnsk\" (UniqueName: \"kubernetes.io/projected/f7c9038f-342a-4d92-b283-086b6a0c7e7e-kube-api-access-5bnsk\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129244 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129262 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-logs\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.129291 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.130144 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.130296 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-nb\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.130382 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-sb\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.130417 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-logs\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.130443 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-dns-svc\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.130933 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-config\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.134002 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.135165 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.135299 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.136867 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.144446 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j58tp\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-kube-api-access-j58tp\") pod \"glance-default-internal-api-0\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.148022 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bnsk\" (UniqueName: \"kubernetes.io/projected/f7c9038f-342a-4d92-b283-086b6a0c7e7e-kube-api-access-5bnsk\") pod \"dnsmasq-dns-784c8875b9-kv9rb\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.206533 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.380735 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:06 crc kubenswrapper[4935]: W1005 08:50:06.388204 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2eca9ca9_d116_47c8_a71b_d8b4ba982c53.slice/crio-8586a3165cc584b4ea7824547cacc845f7b19f48e5086e6c64e4c4dbf96cb57c WatchSource:0}: Error finding container 8586a3165cc584b4ea7824547cacc845f7b19f48e5086e6c64e4c4dbf96cb57c: Status 404 returned error can't find the container with id 8586a3165cc584b4ea7824547cacc845f7b19f48e5086e6c64e4c4dbf96cb57c Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.432766 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:06 crc kubenswrapper[4935]: I1005 08:50:06.696479 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784c8875b9-kv9rb"] Oct 05 08:50:07 crc kubenswrapper[4935]: I1005 08:50:06.999777 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:07 crc kubenswrapper[4935]: W1005 08:50:07.022049 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3381bd4c_8260_420d_ad59_5033cad3815f.slice/crio-f7f9499b691e22fb7d12abacc498b707be651d1633d548d000bbd36d93961096 WatchSource:0}: Error finding container f7f9499b691e22fb7d12abacc498b707be651d1633d548d000bbd36d93961096: Status 404 returned error can't find the container with id f7f9499b691e22fb7d12abacc498b707be651d1633d548d000bbd36d93961096 Oct 05 08:50:07 crc kubenswrapper[4935]: I1005 08:50:07.089489 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:07 crc kubenswrapper[4935]: I1005 08:50:07.317437 4935 generic.go:334] "Generic (PLEG): container finished" podID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" containerID="3f0648edb391492891f6ff2863d71c4842a472070e63491cae88ba5332fc838b" exitCode=0 Oct 05 08:50:07 crc kubenswrapper[4935]: I1005 08:50:07.317933 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" event={"ID":"f7c9038f-342a-4d92-b283-086b6a0c7e7e","Type":"ContainerDied","Data":"3f0648edb391492891f6ff2863d71c4842a472070e63491cae88ba5332fc838b"} Oct 05 08:50:07 crc kubenswrapper[4935]: I1005 08:50:07.318027 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" event={"ID":"f7c9038f-342a-4d92-b283-086b6a0c7e7e","Type":"ContainerStarted","Data":"91d1ac77a5a622e05ebd3ca17b420d961d1ca170570ed21c0dbbf86bb8fcdfb0"} Oct 05 08:50:07 crc kubenswrapper[4935]: I1005 08:50:07.328907 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2eca9ca9-d116-47c8-a71b-d8b4ba982c53","Type":"ContainerStarted","Data":"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187"} Oct 05 08:50:07 crc kubenswrapper[4935]: I1005 08:50:07.328946 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2eca9ca9-d116-47c8-a71b-d8b4ba982c53","Type":"ContainerStarted","Data":"8586a3165cc584b4ea7824547cacc845f7b19f48e5086e6c64e4c4dbf96cb57c"} Oct 05 08:50:07 crc kubenswrapper[4935]: I1005 08:50:07.335193 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3381bd4c-8260-420d-ad59-5033cad3815f","Type":"ContainerStarted","Data":"f7f9499b691e22fb7d12abacc498b707be651d1633d548d000bbd36d93961096"} Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.342782 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3381bd4c-8260-420d-ad59-5033cad3815f","Type":"ContainerStarted","Data":"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987"} Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.343519 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3381bd4c-8260-420d-ad59-5033cad3815f","Type":"ContainerStarted","Data":"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a"} Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.346334 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" event={"ID":"f7c9038f-342a-4d92-b283-086b6a0c7e7e","Type":"ContainerStarted","Data":"515c9704ba06700a92c29c105e08674063a2d70d695f4695faab9e839476e127"} Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.346530 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.349007 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2eca9ca9-d116-47c8-a71b-d8b4ba982c53","Type":"ContainerStarted","Data":"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f"} Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.349104 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerName="glance-log" containerID="cri-o://76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187" gracePeriod=30 Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.349128 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerName="glance-httpd" containerID="cri-o://a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f" gracePeriod=30 Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.369818 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.369797979 podStartE2EDuration="3.369797979s" podCreationTimestamp="2025-10-05 08:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:50:08.361024656 +0000 UTC m=+7042.243651186" watchObservedRunningTime="2025-10-05 08:50:08.369797979 +0000 UTC m=+7042.252424449" Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.388237 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" podStartSLOduration=3.388193567 podStartE2EDuration="3.388193567s" podCreationTimestamp="2025-10-05 08:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:50:08.387365655 +0000 UTC m=+7042.269992155" watchObservedRunningTime="2025-10-05 08:50:08.388193567 +0000 UTC m=+7042.270820057" Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.417664 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.417640178 podStartE2EDuration="3.417640178s" podCreationTimestamp="2025-10-05 08:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:50:08.411021333 +0000 UTC m=+7042.293647813" watchObservedRunningTime="2025-10-05 08:50:08.417640178 +0000 UTC m=+7042.300266638" Oct 05 08:50:08 crc kubenswrapper[4935]: I1005 08:50:08.952668 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.088466 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-config-data\") pod \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.089244 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-combined-ca-bundle\") pod \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.089313 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-ceph\") pod \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.089434 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvgh6\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-kube-api-access-zvgh6\") pod \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.089470 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-httpd-run\") pod \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.089518 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-logs\") pod \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.089547 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-scripts\") pod \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\" (UID: \"2eca9ca9-d116-47c8-a71b-d8b4ba982c53\") " Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.090140 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2eca9ca9-d116-47c8-a71b-d8b4ba982c53" (UID: "2eca9ca9-d116-47c8-a71b-d8b4ba982c53"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.090441 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-logs" (OuterVolumeSpecName: "logs") pod "2eca9ca9-d116-47c8-a71b-d8b4ba982c53" (UID: "2eca9ca9-d116-47c8-a71b-d8b4ba982c53"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.095025 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-scripts" (OuterVolumeSpecName: "scripts") pod "2eca9ca9-d116-47c8-a71b-d8b4ba982c53" (UID: "2eca9ca9-d116-47c8-a71b-d8b4ba982c53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.098690 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-ceph" (OuterVolumeSpecName: "ceph") pod "2eca9ca9-d116-47c8-a71b-d8b4ba982c53" (UID: "2eca9ca9-d116-47c8-a71b-d8b4ba982c53"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.098920 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-kube-api-access-zvgh6" (OuterVolumeSpecName: "kube-api-access-zvgh6") pod "2eca9ca9-d116-47c8-a71b-d8b4ba982c53" (UID: "2eca9ca9-d116-47c8-a71b-d8b4ba982c53"). InnerVolumeSpecName "kube-api-access-zvgh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.123043 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2eca9ca9-d116-47c8-a71b-d8b4ba982c53" (UID: "2eca9ca9-d116-47c8-a71b-d8b4ba982c53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.145037 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-config-data" (OuterVolumeSpecName: "config-data") pod "2eca9ca9-d116-47c8-a71b-d8b4ba982c53" (UID: "2eca9ca9-d116-47c8-a71b-d8b4ba982c53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.191555 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.191598 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.191615 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.191628 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvgh6\" (UniqueName: \"kubernetes.io/projected/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-kube-api-access-zvgh6\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.191642 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.191656 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.191667 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eca9ca9-d116-47c8-a71b-d8b4ba982c53-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.360257 4935 generic.go:334] "Generic (PLEG): container finished" podID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerID="a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f" exitCode=0 Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.360297 4935 generic.go:334] "Generic (PLEG): container finished" podID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerID="76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187" exitCode=143 Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.361525 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.366084 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2eca9ca9-d116-47c8-a71b-d8b4ba982c53","Type":"ContainerDied","Data":"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f"} Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.366143 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2eca9ca9-d116-47c8-a71b-d8b4ba982c53","Type":"ContainerDied","Data":"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187"} Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.366159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2eca9ca9-d116-47c8-a71b-d8b4ba982c53","Type":"ContainerDied","Data":"8586a3165cc584b4ea7824547cacc845f7b19f48e5086e6c64e4c4dbf96cb57c"} Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.366178 4935 scope.go:117] "RemoveContainer" containerID="a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.409911 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.416767 4935 scope.go:117] "RemoveContainer" containerID="76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.422401 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.444298 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:09 crc kubenswrapper[4935]: E1005 08:50:09.444709 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerName="glance-httpd" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.444726 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerName="glance-httpd" Oct 05 08:50:09 crc kubenswrapper[4935]: E1005 08:50:09.444739 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerName="glance-log" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.444746 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerName="glance-log" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.444929 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerName="glance-httpd" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.444958 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" containerName="glance-log" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.445837 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.449104 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.457114 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.463055 4935 scope.go:117] "RemoveContainer" containerID="a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f" Oct 05 08:50:09 crc kubenswrapper[4935]: E1005 08:50:09.463520 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f\": container with ID starting with a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f not found: ID does not exist" containerID="a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.463581 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f"} err="failed to get container status \"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f\": rpc error: code = NotFound desc = could not find container \"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f\": container with ID starting with a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f not found: ID does not exist" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.463609 4935 scope.go:117] "RemoveContainer" containerID="76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187" Oct 05 08:50:09 crc kubenswrapper[4935]: E1005 08:50:09.464078 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187\": container with ID starting with 76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187 not found: ID does not exist" containerID="76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.464099 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187"} err="failed to get container status \"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187\": rpc error: code = NotFound desc = could not find container \"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187\": container with ID starting with 76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187 not found: ID does not exist" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.464112 4935 scope.go:117] "RemoveContainer" containerID="a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.469614 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f"} err="failed to get container status \"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f\": rpc error: code = NotFound desc = could not find container \"a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f\": container with ID starting with a744f073352eb138a4610c94b9f4b391123ad113fe1ff7c9f8a8e4bb8e9bc04f not found: ID does not exist" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.469659 4935 scope.go:117] "RemoveContainer" containerID="76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.472335 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187"} err="failed to get container status \"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187\": rpc error: code = NotFound desc = could not find container \"76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187\": container with ID starting with 76c131909785886725b726f8c9baa9300110e406e18c542ed763941b143b2187 not found: ID does not exist" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.519057 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.598795 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-logs\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.598853 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.599479 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-ceph\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.599519 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.599582 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmdf4\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-kube-api-access-hmdf4\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.599691 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.599749 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.701681 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.701767 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.701871 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-logs\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.702040 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.702202 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-ceph\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.702275 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.702197 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.702387 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmdf4\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-kube-api-access-hmdf4\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.702529 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-logs\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.708018 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.708350 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.709909 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.714223 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-ceph\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.724407 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmdf4\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-kube-api-access-hmdf4\") pod \"glance-default-external-api-0\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " pod="openstack/glance-default-external-api-0" Oct 05 08:50:09 crc kubenswrapper[4935]: I1005 08:50:09.783924 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:50:10 crc kubenswrapper[4935]: I1005 08:50:10.358006 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:50:10 crc kubenswrapper[4935]: W1005 08:50:10.359431 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b8d7f12_35c2_47f7_b6a5_d4699b0467c9.slice/crio-6a389b5c8a7e72358b6d68406c69ceb1849df6ae6e78ef162dcd9d7e110e9816 WatchSource:0}: Error finding container 6a389b5c8a7e72358b6d68406c69ceb1849df6ae6e78ef162dcd9d7e110e9816: Status 404 returned error can't find the container with id 6a389b5c8a7e72358b6d68406c69ceb1849df6ae6e78ef162dcd9d7e110e9816 Oct 05 08:50:10 crc kubenswrapper[4935]: I1005 08:50:10.396689 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9","Type":"ContainerStarted","Data":"6a389b5c8a7e72358b6d68406c69ceb1849df6ae6e78ef162dcd9d7e110e9816"} Oct 05 08:50:10 crc kubenswrapper[4935]: I1005 08:50:10.398603 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" containerName="glance-log" containerID="cri-o://e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a" gracePeriod=30 Oct 05 08:50:10 crc kubenswrapper[4935]: I1005 08:50:10.398711 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" containerName="glance-httpd" containerID="cri-o://af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987" gracePeriod=30 Oct 05 08:50:10 crc kubenswrapper[4935]: I1005 08:50:10.797635 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eca9ca9-d116-47c8-a71b-d8b4ba982c53" path="/var/lib/kubelet/pods/2eca9ca9-d116-47c8-a71b-d8b4ba982c53/volumes" Oct 05 08:50:10 crc kubenswrapper[4935]: I1005 08:50:10.959170 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.130525 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-scripts\") pod \"3381bd4c-8260-420d-ad59-5033cad3815f\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.130598 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-logs\") pod \"3381bd4c-8260-420d-ad59-5033cad3815f\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.130686 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j58tp\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-kube-api-access-j58tp\") pod \"3381bd4c-8260-420d-ad59-5033cad3815f\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.130782 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-ceph\") pod \"3381bd4c-8260-420d-ad59-5033cad3815f\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.130857 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-combined-ca-bundle\") pod \"3381bd4c-8260-420d-ad59-5033cad3815f\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.131275 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-logs" (OuterVolumeSpecName: "logs") pod "3381bd4c-8260-420d-ad59-5033cad3815f" (UID: "3381bd4c-8260-420d-ad59-5033cad3815f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.131648 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-config-data\") pod \"3381bd4c-8260-420d-ad59-5033cad3815f\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.131697 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-httpd-run\") pod \"3381bd4c-8260-420d-ad59-5033cad3815f\" (UID: \"3381bd4c-8260-420d-ad59-5033cad3815f\") " Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.132036 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3381bd4c-8260-420d-ad59-5033cad3815f" (UID: "3381bd4c-8260-420d-ad59-5033cad3815f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.132586 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.132604 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3381bd4c-8260-420d-ad59-5033cad3815f-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.134523 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-kube-api-access-j58tp" (OuterVolumeSpecName: "kube-api-access-j58tp") pod "3381bd4c-8260-420d-ad59-5033cad3815f" (UID: "3381bd4c-8260-420d-ad59-5033cad3815f"). InnerVolumeSpecName "kube-api-access-j58tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.134876 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-ceph" (OuterVolumeSpecName: "ceph") pod "3381bd4c-8260-420d-ad59-5033cad3815f" (UID: "3381bd4c-8260-420d-ad59-5033cad3815f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.135916 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-scripts" (OuterVolumeSpecName: "scripts") pod "3381bd4c-8260-420d-ad59-5033cad3815f" (UID: "3381bd4c-8260-420d-ad59-5033cad3815f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.153575 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3381bd4c-8260-420d-ad59-5033cad3815f" (UID: "3381bd4c-8260-420d-ad59-5033cad3815f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.188717 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-config-data" (OuterVolumeSpecName: "config-data") pod "3381bd4c-8260-420d-ad59-5033cad3815f" (UID: "3381bd4c-8260-420d-ad59-5033cad3815f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.234481 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.234521 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.234534 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j58tp\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-kube-api-access-j58tp\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.234549 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3381bd4c-8260-420d-ad59-5033cad3815f-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.234561 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3381bd4c-8260-420d-ad59-5033cad3815f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.409067 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9","Type":"ContainerStarted","Data":"c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385"} Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.409115 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9","Type":"ContainerStarted","Data":"9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb"} Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.414042 4935 generic.go:334] "Generic (PLEG): container finished" podID="3381bd4c-8260-420d-ad59-5033cad3815f" containerID="af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987" exitCode=0 Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.414097 4935 generic.go:334] "Generic (PLEG): container finished" podID="3381bd4c-8260-420d-ad59-5033cad3815f" containerID="e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a" exitCode=143 Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.414100 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.414105 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3381bd4c-8260-420d-ad59-5033cad3815f","Type":"ContainerDied","Data":"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987"} Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.414229 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3381bd4c-8260-420d-ad59-5033cad3815f","Type":"ContainerDied","Data":"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a"} Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.414241 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3381bd4c-8260-420d-ad59-5033cad3815f","Type":"ContainerDied","Data":"f7f9499b691e22fb7d12abacc498b707be651d1633d548d000bbd36d93961096"} Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.414257 4935 scope.go:117] "RemoveContainer" containerID="af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.440986 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.440964854 podStartE2EDuration="2.440964854s" podCreationTimestamp="2025-10-05 08:50:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:50:11.430051354 +0000 UTC m=+7045.312677824" watchObservedRunningTime="2025-10-05 08:50:11.440964854 +0000 UTC m=+7045.323591314" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.444243 4935 scope.go:117] "RemoveContainer" containerID="e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.470480 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.479395 4935 scope.go:117] "RemoveContainer" containerID="af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987" Oct 05 08:50:11 crc kubenswrapper[4935]: E1005 08:50:11.480120 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987\": container with ID starting with af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987 not found: ID does not exist" containerID="af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.480193 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987"} err="failed to get container status \"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987\": rpc error: code = NotFound desc = could not find container \"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987\": container with ID starting with af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987 not found: ID does not exist" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.480222 4935 scope.go:117] "RemoveContainer" containerID="e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a" Oct 05 08:50:11 crc kubenswrapper[4935]: E1005 08:50:11.480674 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a\": container with ID starting with e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a not found: ID does not exist" containerID="e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.480704 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a"} err="failed to get container status \"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a\": rpc error: code = NotFound desc = could not find container \"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a\": container with ID starting with e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a not found: ID does not exist" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.480724 4935 scope.go:117] "RemoveContainer" containerID="af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.481019 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987"} err="failed to get container status \"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987\": rpc error: code = NotFound desc = could not find container \"af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987\": container with ID starting with af19123f7ab6a4273c7f60117974136ccefbe4881b332c3caea065cfd1a08987 not found: ID does not exist" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.481040 4935 scope.go:117] "RemoveContainer" containerID="e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.481291 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a"} err="failed to get container status \"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a\": rpc error: code = NotFound desc = could not find container \"e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a\": container with ID starting with e124a6d3e91dfe6dfec09e1ad2aa7e434434e548c2b158dddab6c957388e424a not found: ID does not exist" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.483139 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.492942 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:11 crc kubenswrapper[4935]: E1005 08:50:11.493343 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" containerName="glance-log" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.493360 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" containerName="glance-log" Oct 05 08:50:11 crc kubenswrapper[4935]: E1005 08:50:11.493385 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" containerName="glance-httpd" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.493396 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" containerName="glance-httpd" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.493611 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" containerName="glance-log" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.493637 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" containerName="glance-httpd" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.494787 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.496998 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.501223 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.642496 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.642645 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.642682 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.642728 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf69g\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-kube-api-access-kf69g\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.642755 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.642823 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.642861 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-logs\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.744057 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.744172 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.744204 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.744234 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf69g\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-kube-api-access-kf69g\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.744254 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.744280 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.744297 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-logs\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.744749 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-logs\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.746804 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.749914 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.750122 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.751088 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.752712 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.761333 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf69g\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-kube-api-access-kf69g\") pod \"glance-default-internal-api-0\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:50:11 crc kubenswrapper[4935]: I1005 08:50:11.847826 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:12 crc kubenswrapper[4935]: I1005 08:50:12.401024 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:50:12 crc kubenswrapper[4935]: I1005 08:50:12.427271 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"421b8438-b398-4a68-a88b-cf875d82bf1e","Type":"ContainerStarted","Data":"d182e4fc3338243dc6f6313ef64eea424d7dc9204379bec9285ca0895442cea8"} Oct 05 08:50:12 crc kubenswrapper[4935]: I1005 08:50:12.789807 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3381bd4c-8260-420d-ad59-5033cad3815f" path="/var/lib/kubelet/pods/3381bd4c-8260-420d-ad59-5033cad3815f/volumes" Oct 05 08:50:13 crc kubenswrapper[4935]: I1005 08:50:13.439155 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"421b8438-b398-4a68-a88b-cf875d82bf1e","Type":"ContainerStarted","Data":"7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b"} Oct 05 08:50:13 crc kubenswrapper[4935]: I1005 08:50:13.439554 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"421b8438-b398-4a68-a88b-cf875d82bf1e","Type":"ContainerStarted","Data":"62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613"} Oct 05 08:50:13 crc kubenswrapper[4935]: I1005 08:50:13.460807 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.460785371 podStartE2EDuration="2.460785371s" podCreationTimestamp="2025-10-05 08:50:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:50:13.455857821 +0000 UTC m=+7047.338484301" watchObservedRunningTime="2025-10-05 08:50:13.460785371 +0000 UTC m=+7047.343411841" Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.209234 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.294266 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8576ff4fd9-wl24b"] Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.294944 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" podUID="b3d3be9b-6d3c-4174-96f9-7e975624879b" containerName="dnsmasq-dns" containerID="cri-o://af49540a5f86e4963231c61955cc5c74b72559c01cc2534a0e06874ca33fe137" gracePeriod=10 Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.468859 4935 generic.go:334] "Generic (PLEG): container finished" podID="b3d3be9b-6d3c-4174-96f9-7e975624879b" containerID="af49540a5f86e4963231c61955cc5c74b72559c01cc2534a0e06874ca33fe137" exitCode=0 Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.468908 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" event={"ID":"b3d3be9b-6d3c-4174-96f9-7e975624879b","Type":"ContainerDied","Data":"af49540a5f86e4963231c61955cc5c74b72559c01cc2534a0e06874ca33fe137"} Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.788789 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.937328 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-config\") pod \"b3d3be9b-6d3c-4174-96f9-7e975624879b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.937392 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-sb\") pod \"b3d3be9b-6d3c-4174-96f9-7e975624879b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.937466 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-nb\") pod \"b3d3be9b-6d3c-4174-96f9-7e975624879b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.937515 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt2lc\" (UniqueName: \"kubernetes.io/projected/b3d3be9b-6d3c-4174-96f9-7e975624879b-kube-api-access-jt2lc\") pod \"b3d3be9b-6d3c-4174-96f9-7e975624879b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.937541 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-dns-svc\") pod \"b3d3be9b-6d3c-4174-96f9-7e975624879b\" (UID: \"b3d3be9b-6d3c-4174-96f9-7e975624879b\") " Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.951483 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d3be9b-6d3c-4174-96f9-7e975624879b-kube-api-access-jt2lc" (OuterVolumeSpecName: "kube-api-access-jt2lc") pod "b3d3be9b-6d3c-4174-96f9-7e975624879b" (UID: "b3d3be9b-6d3c-4174-96f9-7e975624879b"). InnerVolumeSpecName "kube-api-access-jt2lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.983588 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3d3be9b-6d3c-4174-96f9-7e975624879b" (UID: "b3d3be9b-6d3c-4174-96f9-7e975624879b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.984621 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3d3be9b-6d3c-4174-96f9-7e975624879b" (UID: "b3d3be9b-6d3c-4174-96f9-7e975624879b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.987279 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3d3be9b-6d3c-4174-96f9-7e975624879b" (UID: "b3d3be9b-6d3c-4174-96f9-7e975624879b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:50:16 crc kubenswrapper[4935]: I1005 08:50:16.989228 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-config" (OuterVolumeSpecName: "config") pod "b3d3be9b-6d3c-4174-96f9-7e975624879b" (UID: "b3d3be9b-6d3c-4174-96f9-7e975624879b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.040420 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.040463 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt2lc\" (UniqueName: \"kubernetes.io/projected/b3d3be9b-6d3c-4174-96f9-7e975624879b-kube-api-access-jt2lc\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.040475 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.040485 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.040495 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3d3be9b-6d3c-4174-96f9-7e975624879b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.486369 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" event={"ID":"b3d3be9b-6d3c-4174-96f9-7e975624879b","Type":"ContainerDied","Data":"70f90adec5ae6376cd8bbfb6eeca6cc03dfa9a14f290a1dad80359b729fdbd3e"} Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.486479 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8576ff4fd9-wl24b" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.486499 4935 scope.go:117] "RemoveContainer" containerID="af49540a5f86e4963231c61955cc5c74b72559c01cc2534a0e06874ca33fe137" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.510630 4935 scope.go:117] "RemoveContainer" containerID="b37606a2cedf01304a3c640f175efdff0be220d9a7e53715a94f96416f94100d" Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.527491 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8576ff4fd9-wl24b"] Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.535872 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8576ff4fd9-wl24b"] Oct 05 08:50:17 crc kubenswrapper[4935]: I1005 08:50:17.777787 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:50:17 crc kubenswrapper[4935]: E1005 08:50:17.778338 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:50:18 crc kubenswrapper[4935]: I1005 08:50:18.794896 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d3be9b-6d3c-4174-96f9-7e975624879b" path="/var/lib/kubelet/pods/b3d3be9b-6d3c-4174-96f9-7e975624879b/volumes" Oct 05 08:50:19 crc kubenswrapper[4935]: I1005 08:50:19.785145 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 08:50:19 crc kubenswrapper[4935]: I1005 08:50:19.785528 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 08:50:19 crc kubenswrapper[4935]: I1005 08:50:19.843546 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 08:50:19 crc kubenswrapper[4935]: I1005 08:50:19.844294 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 08:50:20 crc kubenswrapper[4935]: I1005 08:50:20.518376 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 08:50:20 crc kubenswrapper[4935]: I1005 08:50:20.518575 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 08:50:21 crc kubenswrapper[4935]: I1005 08:50:21.848387 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:21 crc kubenswrapper[4935]: I1005 08:50:21.850097 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:21 crc kubenswrapper[4935]: I1005 08:50:21.880011 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:21 crc kubenswrapper[4935]: I1005 08:50:21.894156 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:22 crc kubenswrapper[4935]: I1005 08:50:22.431535 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 08:50:22 crc kubenswrapper[4935]: I1005 08:50:22.435238 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 08:50:22 crc kubenswrapper[4935]: I1005 08:50:22.538855 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:22 crc kubenswrapper[4935]: I1005 08:50:22.538980 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:24 crc kubenswrapper[4935]: I1005 08:50:24.379322 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:24 crc kubenswrapper[4935]: I1005 08:50:24.383102 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 08:50:30 crc kubenswrapper[4935]: I1005 08:50:30.778637 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:50:30 crc kubenswrapper[4935]: E1005 08:50:30.779384 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:50:34 crc kubenswrapper[4935]: I1005 08:50:34.807544 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-rjlzp"] Oct 05 08:50:34 crc kubenswrapper[4935]: E1005 08:50:34.808675 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d3be9b-6d3c-4174-96f9-7e975624879b" containerName="init" Oct 05 08:50:34 crc kubenswrapper[4935]: I1005 08:50:34.808692 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d3be9b-6d3c-4174-96f9-7e975624879b" containerName="init" Oct 05 08:50:34 crc kubenswrapper[4935]: E1005 08:50:34.808712 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d3be9b-6d3c-4174-96f9-7e975624879b" containerName="dnsmasq-dns" Oct 05 08:50:34 crc kubenswrapper[4935]: I1005 08:50:34.808721 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d3be9b-6d3c-4174-96f9-7e975624879b" containerName="dnsmasq-dns" Oct 05 08:50:34 crc kubenswrapper[4935]: I1005 08:50:34.808985 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d3be9b-6d3c-4174-96f9-7e975624879b" containerName="dnsmasq-dns" Oct 05 08:50:34 crc kubenswrapper[4935]: I1005 08:50:34.809778 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rjlzp" Oct 05 08:50:34 crc kubenswrapper[4935]: I1005 08:50:34.820358 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rjlzp"] Oct 05 08:50:34 crc kubenswrapper[4935]: I1005 08:50:34.890217 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxwv2\" (UniqueName: \"kubernetes.io/projected/7c82d411-42ad-40d0-a507-8805af16350f-kube-api-access-cxwv2\") pod \"placement-db-create-rjlzp\" (UID: \"7c82d411-42ad-40d0-a507-8805af16350f\") " pod="openstack/placement-db-create-rjlzp" Oct 05 08:50:34 crc kubenswrapper[4935]: I1005 08:50:34.992274 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxwv2\" (UniqueName: \"kubernetes.io/projected/7c82d411-42ad-40d0-a507-8805af16350f-kube-api-access-cxwv2\") pod \"placement-db-create-rjlzp\" (UID: \"7c82d411-42ad-40d0-a507-8805af16350f\") " pod="openstack/placement-db-create-rjlzp" Oct 05 08:50:35 crc kubenswrapper[4935]: I1005 08:50:35.014138 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxwv2\" (UniqueName: \"kubernetes.io/projected/7c82d411-42ad-40d0-a507-8805af16350f-kube-api-access-cxwv2\") pod \"placement-db-create-rjlzp\" (UID: \"7c82d411-42ad-40d0-a507-8805af16350f\") " pod="openstack/placement-db-create-rjlzp" Oct 05 08:50:35 crc kubenswrapper[4935]: I1005 08:50:35.140212 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rjlzp" Oct 05 08:50:35 crc kubenswrapper[4935]: I1005 08:50:35.657435 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rjlzp"] Oct 05 08:50:36 crc kubenswrapper[4935]: I1005 08:50:36.692063 4935 generic.go:334] "Generic (PLEG): container finished" podID="7c82d411-42ad-40d0-a507-8805af16350f" containerID="7207f2634bb3213d44cec07ddb4f76cf1d2050277444e2b079c788b06155dc5f" exitCode=0 Oct 05 08:50:36 crc kubenswrapper[4935]: I1005 08:50:36.692237 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rjlzp" event={"ID":"7c82d411-42ad-40d0-a507-8805af16350f","Type":"ContainerDied","Data":"7207f2634bb3213d44cec07ddb4f76cf1d2050277444e2b079c788b06155dc5f"} Oct 05 08:50:36 crc kubenswrapper[4935]: I1005 08:50:36.692557 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rjlzp" event={"ID":"7c82d411-42ad-40d0-a507-8805af16350f","Type":"ContainerStarted","Data":"8a757b5e0b2654216a05d08403862c0ecdfdb828bf2eac8e5966ede315a068b1"} Oct 05 08:50:38 crc kubenswrapper[4935]: I1005 08:50:38.165956 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rjlzp" Oct 05 08:50:38 crc kubenswrapper[4935]: I1005 08:50:38.327156 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxwv2\" (UniqueName: \"kubernetes.io/projected/7c82d411-42ad-40d0-a507-8805af16350f-kube-api-access-cxwv2\") pod \"7c82d411-42ad-40d0-a507-8805af16350f\" (UID: \"7c82d411-42ad-40d0-a507-8805af16350f\") " Oct 05 08:50:38 crc kubenswrapper[4935]: I1005 08:50:38.333322 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c82d411-42ad-40d0-a507-8805af16350f-kube-api-access-cxwv2" (OuterVolumeSpecName: "kube-api-access-cxwv2") pod "7c82d411-42ad-40d0-a507-8805af16350f" (UID: "7c82d411-42ad-40d0-a507-8805af16350f"). InnerVolumeSpecName "kube-api-access-cxwv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:38 crc kubenswrapper[4935]: I1005 08:50:38.430441 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxwv2\" (UniqueName: \"kubernetes.io/projected/7c82d411-42ad-40d0-a507-8805af16350f-kube-api-access-cxwv2\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:38 crc kubenswrapper[4935]: I1005 08:50:38.722464 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rjlzp" event={"ID":"7c82d411-42ad-40d0-a507-8805af16350f","Type":"ContainerDied","Data":"8a757b5e0b2654216a05d08403862c0ecdfdb828bf2eac8e5966ede315a068b1"} Oct 05 08:50:38 crc kubenswrapper[4935]: I1005 08:50:38.722538 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a757b5e0b2654216a05d08403862c0ecdfdb828bf2eac8e5966ede315a068b1" Oct 05 08:50:38 crc kubenswrapper[4935]: I1005 08:50:38.722553 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rjlzp" Oct 05 08:50:44 crc kubenswrapper[4935]: I1005 08:50:44.777751 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:50:44 crc kubenswrapper[4935]: E1005 08:50:44.779008 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:50:44 crc kubenswrapper[4935]: I1005 08:50:44.955700 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-acdd-account-create-8qtbg"] Oct 05 08:50:44 crc kubenswrapper[4935]: E1005 08:50:44.956063 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c82d411-42ad-40d0-a507-8805af16350f" containerName="mariadb-database-create" Oct 05 08:50:44 crc kubenswrapper[4935]: I1005 08:50:44.956085 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c82d411-42ad-40d0-a507-8805af16350f" containerName="mariadb-database-create" Oct 05 08:50:44 crc kubenswrapper[4935]: I1005 08:50:44.956298 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c82d411-42ad-40d0-a507-8805af16350f" containerName="mariadb-database-create" Oct 05 08:50:44 crc kubenswrapper[4935]: I1005 08:50:44.956952 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-acdd-account-create-8qtbg" Oct 05 08:50:44 crc kubenswrapper[4935]: I1005 08:50:44.962531 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 05 08:50:44 crc kubenswrapper[4935]: I1005 08:50:44.964389 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-acdd-account-create-8qtbg"] Oct 05 08:50:45 crc kubenswrapper[4935]: I1005 08:50:45.059697 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds8mc\" (UniqueName: \"kubernetes.io/projected/35816157-f3f5-4908-9bb0-0480b2d8970e-kube-api-access-ds8mc\") pod \"placement-acdd-account-create-8qtbg\" (UID: \"35816157-f3f5-4908-9bb0-0480b2d8970e\") " pod="openstack/placement-acdd-account-create-8qtbg" Oct 05 08:50:45 crc kubenswrapper[4935]: I1005 08:50:45.161748 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds8mc\" (UniqueName: \"kubernetes.io/projected/35816157-f3f5-4908-9bb0-0480b2d8970e-kube-api-access-ds8mc\") pod \"placement-acdd-account-create-8qtbg\" (UID: \"35816157-f3f5-4908-9bb0-0480b2d8970e\") " pod="openstack/placement-acdd-account-create-8qtbg" Oct 05 08:50:45 crc kubenswrapper[4935]: I1005 08:50:45.182731 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds8mc\" (UniqueName: \"kubernetes.io/projected/35816157-f3f5-4908-9bb0-0480b2d8970e-kube-api-access-ds8mc\") pod \"placement-acdd-account-create-8qtbg\" (UID: \"35816157-f3f5-4908-9bb0-0480b2d8970e\") " pod="openstack/placement-acdd-account-create-8qtbg" Oct 05 08:50:45 crc kubenswrapper[4935]: I1005 08:50:45.275716 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-acdd-account-create-8qtbg" Oct 05 08:50:45 crc kubenswrapper[4935]: I1005 08:50:45.807637 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-acdd-account-create-8qtbg"] Oct 05 08:50:46 crc kubenswrapper[4935]: I1005 08:50:46.815844 4935 generic.go:334] "Generic (PLEG): container finished" podID="35816157-f3f5-4908-9bb0-0480b2d8970e" containerID="42a1cf7e66fb4fe2e77524328722fcb1608733769fcbcd0f661d15c22748c0ad" exitCode=0 Oct 05 08:50:46 crc kubenswrapper[4935]: I1005 08:50:46.815913 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-acdd-account-create-8qtbg" event={"ID":"35816157-f3f5-4908-9bb0-0480b2d8970e","Type":"ContainerDied","Data":"42a1cf7e66fb4fe2e77524328722fcb1608733769fcbcd0f661d15c22748c0ad"} Oct 05 08:50:46 crc kubenswrapper[4935]: I1005 08:50:46.816216 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-acdd-account-create-8qtbg" event={"ID":"35816157-f3f5-4908-9bb0-0480b2d8970e","Type":"ContainerStarted","Data":"0177222830c9019f201fabad0a741ad7400b604ba9bb0679b2b287fa2bf95d8c"} Oct 05 08:50:48 crc kubenswrapper[4935]: I1005 08:50:48.141079 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-acdd-account-create-8qtbg" Oct 05 08:50:48 crc kubenswrapper[4935]: I1005 08:50:48.326763 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds8mc\" (UniqueName: \"kubernetes.io/projected/35816157-f3f5-4908-9bb0-0480b2d8970e-kube-api-access-ds8mc\") pod \"35816157-f3f5-4908-9bb0-0480b2d8970e\" (UID: \"35816157-f3f5-4908-9bb0-0480b2d8970e\") " Oct 05 08:50:48 crc kubenswrapper[4935]: I1005 08:50:48.332130 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35816157-f3f5-4908-9bb0-0480b2d8970e-kube-api-access-ds8mc" (OuterVolumeSpecName: "kube-api-access-ds8mc") pod "35816157-f3f5-4908-9bb0-0480b2d8970e" (UID: "35816157-f3f5-4908-9bb0-0480b2d8970e"). InnerVolumeSpecName "kube-api-access-ds8mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:48 crc kubenswrapper[4935]: I1005 08:50:48.428410 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds8mc\" (UniqueName: \"kubernetes.io/projected/35816157-f3f5-4908-9bb0-0480b2d8970e-kube-api-access-ds8mc\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:48 crc kubenswrapper[4935]: I1005 08:50:48.838249 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-acdd-account-create-8qtbg" event={"ID":"35816157-f3f5-4908-9bb0-0480b2d8970e","Type":"ContainerDied","Data":"0177222830c9019f201fabad0a741ad7400b604ba9bb0679b2b287fa2bf95d8c"} Oct 05 08:50:48 crc kubenswrapper[4935]: I1005 08:50:48.838632 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0177222830c9019f201fabad0a741ad7400b604ba9bb0679b2b287fa2bf95d8c" Oct 05 08:50:48 crc kubenswrapper[4935]: I1005 08:50:48.838345 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-acdd-account-create-8qtbg" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.221604 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d984f9d9c-s6xtd"] Oct 05 08:50:50 crc kubenswrapper[4935]: E1005 08:50:50.222505 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35816157-f3f5-4908-9bb0-0480b2d8970e" containerName="mariadb-account-create" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.222522 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="35816157-f3f5-4908-9bb0-0480b2d8970e" containerName="mariadb-account-create" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.222762 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="35816157-f3f5-4908-9bb0-0480b2d8970e" containerName="mariadb-account-create" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.223689 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.236922 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-xlvg7"] Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.238138 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.242423 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d984f9d9c-s6xtd"] Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.244258 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.244540 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.244587 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-56h6k" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.260993 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xlvg7"] Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.369350 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-config-data\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.369455 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-sb\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.369500 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpmbq\" (UniqueName: \"kubernetes.io/projected/51a5af21-9df5-4364-924f-29386157ddcf-kube-api-access-bpmbq\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.369525 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-scripts\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.369657 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.369850 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-config\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.370015 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p48b\" (UniqueName: \"kubernetes.io/projected/52eec721-ef6b-46f8-a750-287f89c552eb-kube-api-access-7p48b\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.370150 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-nb\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.370304 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-combined-ca-bundle\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.370414 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52eec721-ef6b-46f8-a750-287f89c552eb-logs\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.471972 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472040 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-config\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472070 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p48b\" (UniqueName: \"kubernetes.io/projected/52eec721-ef6b-46f8-a750-287f89c552eb-kube-api-access-7p48b\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472095 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-nb\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472125 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-combined-ca-bundle\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472149 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52eec721-ef6b-46f8-a750-287f89c552eb-logs\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472186 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-config-data\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472217 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-sb\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472238 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpmbq\" (UniqueName: \"kubernetes.io/projected/51a5af21-9df5-4364-924f-29386157ddcf-kube-api-access-bpmbq\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472255 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-scripts\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.472862 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52eec721-ef6b-46f8-a750-287f89c552eb-logs\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.473224 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.473580 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-nb\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.473750 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-config\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.473862 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-sb\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.477777 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-combined-ca-bundle\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.483504 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-config-data\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.485539 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-scripts\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.488096 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p48b\" (UniqueName: \"kubernetes.io/projected/52eec721-ef6b-46f8-a750-287f89c552eb-kube-api-access-7p48b\") pod \"placement-db-sync-xlvg7\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.489152 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpmbq\" (UniqueName: \"kubernetes.io/projected/51a5af21-9df5-4364-924f-29386157ddcf-kube-api-access-bpmbq\") pod \"dnsmasq-dns-6d984f9d9c-s6xtd\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.556609 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.573435 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.827349 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d984f9d9c-s6xtd"] Oct 05 08:50:50 crc kubenswrapper[4935]: W1005 08:50:50.833786 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a5af21_9df5_4364_924f_29386157ddcf.slice/crio-ccd70b6826c0a988e2eac0d448274386c8c0de999f584db398719aac8b6ee280 WatchSource:0}: Error finding container ccd70b6826c0a988e2eac0d448274386c8c0de999f584db398719aac8b6ee280: Status 404 returned error can't find the container with id ccd70b6826c0a988e2eac0d448274386c8c0de999f584db398719aac8b6ee280 Oct 05 08:50:50 crc kubenswrapper[4935]: I1005 08:50:50.869718 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" event={"ID":"51a5af21-9df5-4364-924f-29386157ddcf","Type":"ContainerStarted","Data":"ccd70b6826c0a988e2eac0d448274386c8c0de999f584db398719aac8b6ee280"} Oct 05 08:50:51 crc kubenswrapper[4935]: I1005 08:50:51.114746 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xlvg7"] Oct 05 08:50:51 crc kubenswrapper[4935]: W1005 08:50:51.142250 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52eec721_ef6b_46f8_a750_287f89c552eb.slice/crio-211b62d72abafda267054b554bc8ae6482e7b01e9495eeb7a388ce1b090563bc WatchSource:0}: Error finding container 211b62d72abafda267054b554bc8ae6482e7b01e9495eeb7a388ce1b090563bc: Status 404 returned error can't find the container with id 211b62d72abafda267054b554bc8ae6482e7b01e9495eeb7a388ce1b090563bc Oct 05 08:50:51 crc kubenswrapper[4935]: I1005 08:50:51.879524 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xlvg7" event={"ID":"52eec721-ef6b-46f8-a750-287f89c552eb","Type":"ContainerStarted","Data":"211b62d72abafda267054b554bc8ae6482e7b01e9495eeb7a388ce1b090563bc"} Oct 05 08:50:51 crc kubenswrapper[4935]: I1005 08:50:51.881666 4935 generic.go:334] "Generic (PLEG): container finished" podID="51a5af21-9df5-4364-924f-29386157ddcf" containerID="220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8" exitCode=0 Oct 05 08:50:51 crc kubenswrapper[4935]: I1005 08:50:51.881776 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" event={"ID":"51a5af21-9df5-4364-924f-29386157ddcf","Type":"ContainerDied","Data":"220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8"} Oct 05 08:50:52 crc kubenswrapper[4935]: I1005 08:50:52.896643 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" event={"ID":"51a5af21-9df5-4364-924f-29386157ddcf","Type":"ContainerStarted","Data":"1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317"} Oct 05 08:50:52 crc kubenswrapper[4935]: I1005 08:50:52.896979 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:50:52 crc kubenswrapper[4935]: I1005 08:50:52.919222 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" podStartSLOduration=2.919204903 podStartE2EDuration="2.919204903s" podCreationTimestamp="2025-10-05 08:50:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:50:52.917834537 +0000 UTC m=+7086.800461007" watchObservedRunningTime="2025-10-05 08:50:52.919204903 +0000 UTC m=+7086.801831363" Oct 05 08:50:54 crc kubenswrapper[4935]: I1005 08:50:54.915454 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xlvg7" event={"ID":"52eec721-ef6b-46f8-a750-287f89c552eb","Type":"ContainerStarted","Data":"3e9e1b7c7535382a5b8bbdee4c249266afa59ca2ab5beb23bd1c8446bc01cc9e"} Oct 05 08:50:54 crc kubenswrapper[4935]: I1005 08:50:54.980790 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-xlvg7" podStartSLOduration=2.000771471 podStartE2EDuration="4.980762297s" podCreationTimestamp="2025-10-05 08:50:50 +0000 UTC" firstStartedPulling="2025-10-05 08:50:51.149269014 +0000 UTC m=+7085.031895474" lastFinishedPulling="2025-10-05 08:50:54.12925983 +0000 UTC m=+7088.011886300" observedRunningTime="2025-10-05 08:50:54.970804613 +0000 UTC m=+7088.853431103" watchObservedRunningTime="2025-10-05 08:50:54.980762297 +0000 UTC m=+7088.863388797" Oct 05 08:50:55 crc kubenswrapper[4935]: I1005 08:50:55.929206 4935 generic.go:334] "Generic (PLEG): container finished" podID="52eec721-ef6b-46f8-a750-287f89c552eb" containerID="3e9e1b7c7535382a5b8bbdee4c249266afa59ca2ab5beb23bd1c8446bc01cc9e" exitCode=0 Oct 05 08:50:55 crc kubenswrapper[4935]: I1005 08:50:55.929256 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xlvg7" event={"ID":"52eec721-ef6b-46f8-a750-287f89c552eb","Type":"ContainerDied","Data":"3e9e1b7c7535382a5b8bbdee4c249266afa59ca2ab5beb23bd1c8446bc01cc9e"} Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.386388 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.504281 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52eec721-ef6b-46f8-a750-287f89c552eb-logs\") pod \"52eec721-ef6b-46f8-a750-287f89c552eb\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.504432 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-combined-ca-bundle\") pod \"52eec721-ef6b-46f8-a750-287f89c552eb\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.504482 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p48b\" (UniqueName: \"kubernetes.io/projected/52eec721-ef6b-46f8-a750-287f89c552eb-kube-api-access-7p48b\") pod \"52eec721-ef6b-46f8-a750-287f89c552eb\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.504514 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-scripts\") pod \"52eec721-ef6b-46f8-a750-287f89c552eb\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.504679 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-config-data\") pod \"52eec721-ef6b-46f8-a750-287f89c552eb\" (UID: \"52eec721-ef6b-46f8-a750-287f89c552eb\") " Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.504975 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52eec721-ef6b-46f8-a750-287f89c552eb-logs" (OuterVolumeSpecName: "logs") pod "52eec721-ef6b-46f8-a750-287f89c552eb" (UID: "52eec721-ef6b-46f8-a750-287f89c552eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.505431 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52eec721-ef6b-46f8-a750-287f89c552eb-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.511280 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52eec721-ef6b-46f8-a750-287f89c552eb-kube-api-access-7p48b" (OuterVolumeSpecName: "kube-api-access-7p48b") pod "52eec721-ef6b-46f8-a750-287f89c552eb" (UID: "52eec721-ef6b-46f8-a750-287f89c552eb"). InnerVolumeSpecName "kube-api-access-7p48b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.512137 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-scripts" (OuterVolumeSpecName: "scripts") pod "52eec721-ef6b-46f8-a750-287f89c552eb" (UID: "52eec721-ef6b-46f8-a750-287f89c552eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.536989 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-config-data" (OuterVolumeSpecName: "config-data") pod "52eec721-ef6b-46f8-a750-287f89c552eb" (UID: "52eec721-ef6b-46f8-a750-287f89c552eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.544856 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52eec721-ef6b-46f8-a750-287f89c552eb" (UID: "52eec721-ef6b-46f8-a750-287f89c552eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.607137 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.607204 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.607237 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p48b\" (UniqueName: \"kubernetes.io/projected/52eec721-ef6b-46f8-a750-287f89c552eb-kube-api-access-7p48b\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.607267 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52eec721-ef6b-46f8-a750-287f89c552eb-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.960051 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xlvg7" event={"ID":"52eec721-ef6b-46f8-a750-287f89c552eb","Type":"ContainerDied","Data":"211b62d72abafda267054b554bc8ae6482e7b01e9495eeb7a388ce1b090563bc"} Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.960346 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xlvg7" Oct 05 08:50:57 crc kubenswrapper[4935]: I1005 08:50:57.960369 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="211b62d72abafda267054b554bc8ae6482e7b01e9495eeb7a388ce1b090563bc" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.509621 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7bd4bdf8fb-4xbs4"] Oct 05 08:50:58 crc kubenswrapper[4935]: E1005 08:50:58.510175 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52eec721-ef6b-46f8-a750-287f89c552eb" containerName="placement-db-sync" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.510197 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="52eec721-ef6b-46f8-a750-287f89c552eb" containerName="placement-db-sync" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.510552 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="52eec721-ef6b-46f8-a750-287f89c552eb" containerName="placement-db-sync" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.512175 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.515829 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.516406 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.516740 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-56h6k" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.530258 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7bd4bdf8fb-4xbs4"] Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.627966 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69c4aba6-a024-4c0a-b8de-213c6119bb73-logs\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.628035 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf54l\" (UniqueName: \"kubernetes.io/projected/69c4aba6-a024-4c0a-b8de-213c6119bb73-kube-api-access-nf54l\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.628077 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-scripts\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.628940 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-combined-ca-bundle\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.629018 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-config-data\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.730677 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69c4aba6-a024-4c0a-b8de-213c6119bb73-logs\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.730744 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf54l\" (UniqueName: \"kubernetes.io/projected/69c4aba6-a024-4c0a-b8de-213c6119bb73-kube-api-access-nf54l\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.730779 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-scripts\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.730865 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-combined-ca-bundle\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.730916 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-config-data\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.731331 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69c4aba6-a024-4c0a-b8de-213c6119bb73-logs\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.738036 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-scripts\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.738457 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-config-data\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.739656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c4aba6-a024-4c0a-b8de-213c6119bb73-combined-ca-bundle\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.758511 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf54l\" (UniqueName: \"kubernetes.io/projected/69c4aba6-a024-4c0a-b8de-213c6119bb73-kube-api-access-nf54l\") pod \"placement-7bd4bdf8fb-4xbs4\" (UID: \"69c4aba6-a024-4c0a-b8de-213c6119bb73\") " pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.777214 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:50:58 crc kubenswrapper[4935]: E1005 08:50:58.777644 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:50:58 crc kubenswrapper[4935]: I1005 08:50:58.838992 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:50:59 crc kubenswrapper[4935]: I1005 08:50:59.133457 4935 scope.go:117] "RemoveContainer" containerID="8d4454891d58dc3205d7d1008a031f209f6a1af65a250b8466835c51e55a4115" Oct 05 08:50:59 crc kubenswrapper[4935]: I1005 08:50:59.164153 4935 scope.go:117] "RemoveContainer" containerID="ae1dab29c65968860231c27e67156bdb9501a55e08cd293fa1e400be513d97f3" Oct 05 08:50:59 crc kubenswrapper[4935]: I1005 08:50:59.292863 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7bd4bdf8fb-4xbs4"] Oct 05 08:50:59 crc kubenswrapper[4935]: I1005 08:50:59.980560 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7bd4bdf8fb-4xbs4" event={"ID":"69c4aba6-a024-4c0a-b8de-213c6119bb73","Type":"ContainerStarted","Data":"8366ebfcf9e626d676234529cc468ec8e215fc4cc012d8edb4c098d9bbb35434"} Oct 05 08:50:59 crc kubenswrapper[4935]: I1005 08:50:59.980915 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7bd4bdf8fb-4xbs4" event={"ID":"69c4aba6-a024-4c0a-b8de-213c6119bb73","Type":"ContainerStarted","Data":"72a6d4cfa9a71b698050e8ab1d81a48637b5dddcf72eafaa5131d1004d9a3a6d"} Oct 05 08:50:59 crc kubenswrapper[4935]: I1005 08:50:59.980930 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7bd4bdf8fb-4xbs4" event={"ID":"69c4aba6-a024-4c0a-b8de-213c6119bb73","Type":"ContainerStarted","Data":"00685f789c32321a3a6cedf2955b64d7715bfeb9e5bad17439ec93323c001c5e"} Oct 05 08:50:59 crc kubenswrapper[4935]: I1005 08:50:59.982306 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:51:00 crc kubenswrapper[4935]: I1005 08:51:00.008680 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7bd4bdf8fb-4xbs4" podStartSLOduration=2.008658795 podStartE2EDuration="2.008658795s" podCreationTimestamp="2025-10-05 08:50:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:51:00.004799653 +0000 UTC m=+7093.887426193" watchObservedRunningTime="2025-10-05 08:51:00.008658795 +0000 UTC m=+7093.891285255" Oct 05 08:51:00 crc kubenswrapper[4935]: I1005 08:51:00.558141 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:51:00 crc kubenswrapper[4935]: I1005 08:51:00.615500 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784c8875b9-kv9rb"] Oct 05 08:51:00 crc kubenswrapper[4935]: I1005 08:51:00.616186 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" podUID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" containerName="dnsmasq-dns" containerID="cri-o://515c9704ba06700a92c29c105e08674063a2d70d695f4695faab9e839476e127" gracePeriod=10 Oct 05 08:51:00 crc kubenswrapper[4935]: I1005 08:51:00.997234 4935 generic.go:334] "Generic (PLEG): container finished" podID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" containerID="515c9704ba06700a92c29c105e08674063a2d70d695f4695faab9e839476e127" exitCode=0 Oct 05 08:51:00 crc kubenswrapper[4935]: I1005 08:51:00.997334 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" event={"ID":"f7c9038f-342a-4d92-b283-086b6a0c7e7e","Type":"ContainerDied","Data":"515c9704ba06700a92c29c105e08674063a2d70d695f4695faab9e839476e127"} Oct 05 08:51:00 crc kubenswrapper[4935]: I1005 08:51:00.997431 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.204138 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.373362 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-sb\") pod \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.373622 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-config\") pod \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.373730 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-nb\") pod \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.373817 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bnsk\" (UniqueName: \"kubernetes.io/projected/f7c9038f-342a-4d92-b283-086b6a0c7e7e-kube-api-access-5bnsk\") pod \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.373949 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-dns-svc\") pod \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\" (UID: \"f7c9038f-342a-4d92-b283-086b6a0c7e7e\") " Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.379427 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c9038f-342a-4d92-b283-086b6a0c7e7e-kube-api-access-5bnsk" (OuterVolumeSpecName: "kube-api-access-5bnsk") pod "f7c9038f-342a-4d92-b283-086b6a0c7e7e" (UID: "f7c9038f-342a-4d92-b283-086b6a0c7e7e"). InnerVolumeSpecName "kube-api-access-5bnsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.420513 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-config" (OuterVolumeSpecName: "config") pod "f7c9038f-342a-4d92-b283-086b6a0c7e7e" (UID: "f7c9038f-342a-4d92-b283-086b6a0c7e7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.432444 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f7c9038f-342a-4d92-b283-086b6a0c7e7e" (UID: "f7c9038f-342a-4d92-b283-086b6a0c7e7e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.436755 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f7c9038f-342a-4d92-b283-086b6a0c7e7e" (UID: "f7c9038f-342a-4d92-b283-086b6a0c7e7e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.442060 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f7c9038f-342a-4d92-b283-086b6a0c7e7e" (UID: "f7c9038f-342a-4d92-b283-086b6a0c7e7e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.476154 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.476202 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.476222 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.476241 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bnsk\" (UniqueName: \"kubernetes.io/projected/f7c9038f-342a-4d92-b283-086b6a0c7e7e-kube-api-access-5bnsk\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:01 crc kubenswrapper[4935]: I1005 08:51:01.476260 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7c9038f-342a-4d92-b283-086b6a0c7e7e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:02 crc kubenswrapper[4935]: I1005 08:51:02.010329 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" Oct 05 08:51:02 crc kubenswrapper[4935]: I1005 08:51:02.010309 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784c8875b9-kv9rb" event={"ID":"f7c9038f-342a-4d92-b283-086b6a0c7e7e","Type":"ContainerDied","Data":"91d1ac77a5a622e05ebd3ca17b420d961d1ca170570ed21c0dbbf86bb8fcdfb0"} Oct 05 08:51:02 crc kubenswrapper[4935]: I1005 08:51:02.010578 4935 scope.go:117] "RemoveContainer" containerID="515c9704ba06700a92c29c105e08674063a2d70d695f4695faab9e839476e127" Oct 05 08:51:02 crc kubenswrapper[4935]: I1005 08:51:02.045427 4935 scope.go:117] "RemoveContainer" containerID="3f0648edb391492891f6ff2863d71c4842a472070e63491cae88ba5332fc838b" Oct 05 08:51:02 crc kubenswrapper[4935]: I1005 08:51:02.054175 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784c8875b9-kv9rb"] Oct 05 08:51:02 crc kubenswrapper[4935]: I1005 08:51:02.061957 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784c8875b9-kv9rb"] Oct 05 08:51:02 crc kubenswrapper[4935]: I1005 08:51:02.789609 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" path="/var/lib/kubelet/pods/f7c9038f-342a-4d92-b283-086b6a0c7e7e/volumes" Oct 05 08:51:11 crc kubenswrapper[4935]: I1005 08:51:11.777927 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:51:11 crc kubenswrapper[4935]: E1005 08:51:11.778780 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:51:24 crc kubenswrapper[4935]: I1005 08:51:24.778862 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:51:24 crc kubenswrapper[4935]: E1005 08:51:24.782745 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:51:29 crc kubenswrapper[4935]: I1005 08:51:29.831951 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:51:29 crc kubenswrapper[4935]: I1005 08:51:29.872984 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7bd4bdf8fb-4xbs4" Oct 05 08:51:35 crc kubenswrapper[4935]: I1005 08:51:35.777976 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:51:35 crc kubenswrapper[4935]: E1005 08:51:35.779163 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:51:48 crc kubenswrapper[4935]: I1005 08:51:48.777447 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:51:48 crc kubenswrapper[4935]: E1005 08:51:48.778562 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.375764 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-m8ddf"] Oct 05 08:51:51 crc kubenswrapper[4935]: E1005 08:51:51.377396 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" containerName="dnsmasq-dns" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.377473 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" containerName="dnsmasq-dns" Oct 05 08:51:51 crc kubenswrapper[4935]: E1005 08:51:51.377536 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" containerName="init" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.377589 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" containerName="init" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.377829 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c9038f-342a-4d92-b283-086b6a0c7e7e" containerName="dnsmasq-dns" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.378505 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m8ddf" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.390025 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-m8ddf"] Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.457742 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-24rg4"] Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.459604 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-24rg4" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.468759 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-24rg4"] Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.476406 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkmk6\" (UniqueName: \"kubernetes.io/projected/183c385d-9a5a-484f-b4b4-b59f41a3af84-kube-api-access-rkmk6\") pod \"nova-api-db-create-m8ddf\" (UID: \"183c385d-9a5a-484f-b4b4-b59f41a3af84\") " pod="openstack/nova-api-db-create-m8ddf" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.565626 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-fl9t6"] Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.566858 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fl9t6" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.571204 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fl9t6"] Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.577523 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkt4x\" (UniqueName: \"kubernetes.io/projected/325ad13c-f69b-4c9d-9053-ad415c1f47a0-kube-api-access-zkt4x\") pod \"nova-cell0-db-create-24rg4\" (UID: \"325ad13c-f69b-4c9d-9053-ad415c1f47a0\") " pod="openstack/nova-cell0-db-create-24rg4" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.577615 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkmk6\" (UniqueName: \"kubernetes.io/projected/183c385d-9a5a-484f-b4b4-b59f41a3af84-kube-api-access-rkmk6\") pod \"nova-api-db-create-m8ddf\" (UID: \"183c385d-9a5a-484f-b4b4-b59f41a3af84\") " pod="openstack/nova-api-db-create-m8ddf" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.600570 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkmk6\" (UniqueName: \"kubernetes.io/projected/183c385d-9a5a-484f-b4b4-b59f41a3af84-kube-api-access-rkmk6\") pod \"nova-api-db-create-m8ddf\" (UID: \"183c385d-9a5a-484f-b4b4-b59f41a3af84\") " pod="openstack/nova-api-db-create-m8ddf" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.679559 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr67h\" (UniqueName: \"kubernetes.io/projected/86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e-kube-api-access-sr67h\") pod \"nova-cell1-db-create-fl9t6\" (UID: \"86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e\") " pod="openstack/nova-cell1-db-create-fl9t6" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.679811 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkt4x\" (UniqueName: \"kubernetes.io/projected/325ad13c-f69b-4c9d-9053-ad415c1f47a0-kube-api-access-zkt4x\") pod \"nova-cell0-db-create-24rg4\" (UID: \"325ad13c-f69b-4c9d-9053-ad415c1f47a0\") " pod="openstack/nova-cell0-db-create-24rg4" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.695495 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkt4x\" (UniqueName: \"kubernetes.io/projected/325ad13c-f69b-4c9d-9053-ad415c1f47a0-kube-api-access-zkt4x\") pod \"nova-cell0-db-create-24rg4\" (UID: \"325ad13c-f69b-4c9d-9053-ad415c1f47a0\") " pod="openstack/nova-cell0-db-create-24rg4" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.700298 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m8ddf" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.779518 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-24rg4" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.780663 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr67h\" (UniqueName: \"kubernetes.io/projected/86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e-kube-api-access-sr67h\") pod \"nova-cell1-db-create-fl9t6\" (UID: \"86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e\") " pod="openstack/nova-cell1-db-create-fl9t6" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.811200 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr67h\" (UniqueName: \"kubernetes.io/projected/86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e-kube-api-access-sr67h\") pod \"nova-cell1-db-create-fl9t6\" (UID: \"86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e\") " pod="openstack/nova-cell1-db-create-fl9t6" Oct 05 08:51:51 crc kubenswrapper[4935]: I1005 08:51:51.889624 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fl9t6" Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.199681 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-m8ddf"] Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.285348 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-24rg4"] Oct 05 08:51:52 crc kubenswrapper[4935]: W1005 08:51:52.293021 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod325ad13c_f69b_4c9d_9053_ad415c1f47a0.slice/crio-2c3a6a305a3ffc81ca5d129ba58541e6fabe2e2e1fd46e76dd7046fa23fa4677 WatchSource:0}: Error finding container 2c3a6a305a3ffc81ca5d129ba58541e6fabe2e2e1fd46e76dd7046fa23fa4677: Status 404 returned error can't find the container with id 2c3a6a305a3ffc81ca5d129ba58541e6fabe2e2e1fd46e76dd7046fa23fa4677 Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.357202 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fl9t6"] Oct 05 08:51:52 crc kubenswrapper[4935]: W1005 08:51:52.361406 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86b83239_95a1_4f6b_b6cf_72aa1bf3cb6e.slice/crio-5375fc083ae0c9ade22d40e70ba808ad114547c8b461495f865f7628db144e02 WatchSource:0}: Error finding container 5375fc083ae0c9ade22d40e70ba808ad114547c8b461495f865f7628db144e02: Status 404 returned error can't find the container with id 5375fc083ae0c9ade22d40e70ba808ad114547c8b461495f865f7628db144e02 Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.520534 4935 generic.go:334] "Generic (PLEG): container finished" podID="183c385d-9a5a-484f-b4b4-b59f41a3af84" containerID="242b46b82049ae686f91fa3c5c4343f37219bc3f3ea5872b48c328caddfe92bc" exitCode=0 Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.520590 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m8ddf" event={"ID":"183c385d-9a5a-484f-b4b4-b59f41a3af84","Type":"ContainerDied","Data":"242b46b82049ae686f91fa3c5c4343f37219bc3f3ea5872b48c328caddfe92bc"} Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.522020 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m8ddf" event={"ID":"183c385d-9a5a-484f-b4b4-b59f41a3af84","Type":"ContainerStarted","Data":"04697002a6296cead9dce435d9fb0cf649676cd96c97bafa33ff22145fdf7182"} Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.523275 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fl9t6" event={"ID":"86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e","Type":"ContainerStarted","Data":"5375fc083ae0c9ade22d40e70ba808ad114547c8b461495f865f7628db144e02"} Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.525074 4935 generic.go:334] "Generic (PLEG): container finished" podID="325ad13c-f69b-4c9d-9053-ad415c1f47a0" containerID="0cbc289339fa8b5c1686b974fdd9fba076ddf5bc6fbb7f9de012a0846900857f" exitCode=0 Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.525148 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-24rg4" event={"ID":"325ad13c-f69b-4c9d-9053-ad415c1f47a0","Type":"ContainerDied","Data":"0cbc289339fa8b5c1686b974fdd9fba076ddf5bc6fbb7f9de012a0846900857f"} Oct 05 08:51:52 crc kubenswrapper[4935]: I1005 08:51:52.525276 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-24rg4" event={"ID":"325ad13c-f69b-4c9d-9053-ad415c1f47a0","Type":"ContainerStarted","Data":"2c3a6a305a3ffc81ca5d129ba58541e6fabe2e2e1fd46e76dd7046fa23fa4677"} Oct 05 08:51:53 crc kubenswrapper[4935]: I1005 08:51:53.543287 4935 generic.go:334] "Generic (PLEG): container finished" podID="86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e" containerID="3a6039b9da6eb593a8bf1904f927686f79a14ab313267d01c5f9232650f89fae" exitCode=0 Oct 05 08:51:53 crc kubenswrapper[4935]: I1005 08:51:53.544218 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fl9t6" event={"ID":"86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e","Type":"ContainerDied","Data":"3a6039b9da6eb593a8bf1904f927686f79a14ab313267d01c5f9232650f89fae"} Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.004647 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-24rg4" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.009670 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m8ddf" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.121939 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkmk6\" (UniqueName: \"kubernetes.io/projected/183c385d-9a5a-484f-b4b4-b59f41a3af84-kube-api-access-rkmk6\") pod \"183c385d-9a5a-484f-b4b4-b59f41a3af84\" (UID: \"183c385d-9a5a-484f-b4b4-b59f41a3af84\") " Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.122144 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkt4x\" (UniqueName: \"kubernetes.io/projected/325ad13c-f69b-4c9d-9053-ad415c1f47a0-kube-api-access-zkt4x\") pod \"325ad13c-f69b-4c9d-9053-ad415c1f47a0\" (UID: \"325ad13c-f69b-4c9d-9053-ad415c1f47a0\") " Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.128107 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/325ad13c-f69b-4c9d-9053-ad415c1f47a0-kube-api-access-zkt4x" (OuterVolumeSpecName: "kube-api-access-zkt4x") pod "325ad13c-f69b-4c9d-9053-ad415c1f47a0" (UID: "325ad13c-f69b-4c9d-9053-ad415c1f47a0"). InnerVolumeSpecName "kube-api-access-zkt4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.128375 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/183c385d-9a5a-484f-b4b4-b59f41a3af84-kube-api-access-rkmk6" (OuterVolumeSpecName: "kube-api-access-rkmk6") pod "183c385d-9a5a-484f-b4b4-b59f41a3af84" (UID: "183c385d-9a5a-484f-b4b4-b59f41a3af84"). InnerVolumeSpecName "kube-api-access-rkmk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.224347 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkmk6\" (UniqueName: \"kubernetes.io/projected/183c385d-9a5a-484f-b4b4-b59f41a3af84-kube-api-access-rkmk6\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.224390 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkt4x\" (UniqueName: \"kubernetes.io/projected/325ad13c-f69b-4c9d-9053-ad415c1f47a0-kube-api-access-zkt4x\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.556317 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-24rg4" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.556340 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-24rg4" event={"ID":"325ad13c-f69b-4c9d-9053-ad415c1f47a0","Type":"ContainerDied","Data":"2c3a6a305a3ffc81ca5d129ba58541e6fabe2e2e1fd46e76dd7046fa23fa4677"} Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.556393 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c3a6a305a3ffc81ca5d129ba58541e6fabe2e2e1fd46e76dd7046fa23fa4677" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.559967 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m8ddf" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.560028 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m8ddf" event={"ID":"183c385d-9a5a-484f-b4b4-b59f41a3af84","Type":"ContainerDied","Data":"04697002a6296cead9dce435d9fb0cf649676cd96c97bafa33ff22145fdf7182"} Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.560075 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04697002a6296cead9dce435d9fb0cf649676cd96c97bafa33ff22145fdf7182" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.819663 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fl9t6" Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.936855 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr67h\" (UniqueName: \"kubernetes.io/projected/86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e-kube-api-access-sr67h\") pod \"86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e\" (UID: \"86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e\") " Oct 05 08:51:54 crc kubenswrapper[4935]: I1005 08:51:54.940185 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e-kube-api-access-sr67h" (OuterVolumeSpecName: "kube-api-access-sr67h") pod "86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e" (UID: "86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e"). InnerVolumeSpecName "kube-api-access-sr67h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:51:55 crc kubenswrapper[4935]: I1005 08:51:55.039242 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr67h\" (UniqueName: \"kubernetes.io/projected/86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e-kube-api-access-sr67h\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:55 crc kubenswrapper[4935]: I1005 08:51:55.572575 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fl9t6" event={"ID":"86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e","Type":"ContainerDied","Data":"5375fc083ae0c9ade22d40e70ba808ad114547c8b461495f865f7628db144e02"} Oct 05 08:51:55 crc kubenswrapper[4935]: I1005 08:51:55.572617 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5375fc083ae0c9ade22d40e70ba808ad114547c8b461495f865f7628db144e02" Oct 05 08:51:55 crc kubenswrapper[4935]: I1005 08:51:55.572627 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fl9t6" Oct 05 08:51:59 crc kubenswrapper[4935]: I1005 08:51:59.277341 4935 scope.go:117] "RemoveContainer" containerID="39d994f461a32cf50f0932edb6149c14f8cc8201b0e968da6a9ea2dfac8f073d" Oct 05 08:51:59 crc kubenswrapper[4935]: I1005 08:51:59.312207 4935 scope.go:117] "RemoveContainer" containerID="581f062ac3b0d74adf16872bb5080539ac36ecd2e35eb50f0df16615faa0bd9c" Oct 05 08:51:59 crc kubenswrapper[4935]: I1005 08:51:59.370481 4935 scope.go:117] "RemoveContainer" containerID="4042722c3c48e582f0e47e231d395266889a4315ca99fc8831403ca27161085c" Oct 05 08:51:59 crc kubenswrapper[4935]: I1005 08:51:59.394859 4935 scope.go:117] "RemoveContainer" containerID="99fb81a8f2a8422521134d63449b963210936354f48562b50f7605c4389011fa" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.620956 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0f81-account-create-brq6v"] Oct 05 08:52:01 crc kubenswrapper[4935]: E1005 08:52:01.627100 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.627389 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: E1005 08:52:01.628048 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="183c385d-9a5a-484f-b4b4-b59f41a3af84" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.628069 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="183c385d-9a5a-484f-b4b4-b59f41a3af84" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: E1005 08:52:01.628121 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325ad13c-f69b-4c9d-9053-ad415c1f47a0" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.628137 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="325ad13c-f69b-4c9d-9053-ad415c1f47a0" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.628527 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="325ad13c-f69b-4c9d-9053-ad415c1f47a0" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.628573 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="183c385d-9a5a-484f-b4b4-b59f41a3af84" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.628588 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e" containerName="mariadb-database-create" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.629797 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0f81-account-create-brq6v" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.634640 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0f81-account-create-brq6v"] Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.635609 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.779425 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:52:01 crc kubenswrapper[4935]: E1005 08:52:01.779714 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.792228 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmgm7\" (UniqueName: \"kubernetes.io/projected/4c993f0b-5b16-4710-89ce-e5e1881479bb-kube-api-access-kmgm7\") pod \"nova-api-0f81-account-create-brq6v\" (UID: \"4c993f0b-5b16-4710-89ce-e5e1881479bb\") " pod="openstack/nova-api-0f81-account-create-brq6v" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.802361 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f522-account-create-rqblm"] Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.804467 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f522-account-create-rqblm" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.807042 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.818343 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f522-account-create-rqblm"] Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.893647 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmgm7\" (UniqueName: \"kubernetes.io/projected/4c993f0b-5b16-4710-89ce-e5e1881479bb-kube-api-access-kmgm7\") pod \"nova-api-0f81-account-create-brq6v\" (UID: \"4c993f0b-5b16-4710-89ce-e5e1881479bb\") " pod="openstack/nova-api-0f81-account-create-brq6v" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.893734 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh6bx\" (UniqueName: \"kubernetes.io/projected/ca9810c5-05cd-440e-8b35-54940b7beb4f-kube-api-access-wh6bx\") pod \"nova-cell0-f522-account-create-rqblm\" (UID: \"ca9810c5-05cd-440e-8b35-54940b7beb4f\") " pod="openstack/nova-cell0-f522-account-create-rqblm" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.926702 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmgm7\" (UniqueName: \"kubernetes.io/projected/4c993f0b-5b16-4710-89ce-e5e1881479bb-kube-api-access-kmgm7\") pod \"nova-api-0f81-account-create-brq6v\" (UID: \"4c993f0b-5b16-4710-89ce-e5e1881479bb\") " pod="openstack/nova-api-0f81-account-create-brq6v" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.978129 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0f81-account-create-brq6v" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.995242 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh6bx\" (UniqueName: \"kubernetes.io/projected/ca9810c5-05cd-440e-8b35-54940b7beb4f-kube-api-access-wh6bx\") pod \"nova-cell0-f522-account-create-rqblm\" (UID: \"ca9810c5-05cd-440e-8b35-54940b7beb4f\") " pod="openstack/nova-cell0-f522-account-create-rqblm" Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.997069 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-83b2-account-create-5mvxp"] Oct 05 08:52:01 crc kubenswrapper[4935]: I1005 08:52:01.998800 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-83b2-account-create-5mvxp" Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.001665 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.014820 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh6bx\" (UniqueName: \"kubernetes.io/projected/ca9810c5-05cd-440e-8b35-54940b7beb4f-kube-api-access-wh6bx\") pod \"nova-cell0-f522-account-create-rqblm\" (UID: \"ca9810c5-05cd-440e-8b35-54940b7beb4f\") " pod="openstack/nova-cell0-f522-account-create-rqblm" Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.019669 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-83b2-account-create-5mvxp"] Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.097047 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj55h\" (UniqueName: \"kubernetes.io/projected/0f71a336-51f3-4ad9-8dab-bc801ac5c0d9-kube-api-access-pj55h\") pod \"nova-cell1-83b2-account-create-5mvxp\" (UID: \"0f71a336-51f3-4ad9-8dab-bc801ac5c0d9\") " pod="openstack/nova-cell1-83b2-account-create-5mvxp" Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.121337 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f522-account-create-rqblm" Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.199625 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj55h\" (UniqueName: \"kubernetes.io/projected/0f71a336-51f3-4ad9-8dab-bc801ac5c0d9-kube-api-access-pj55h\") pod \"nova-cell1-83b2-account-create-5mvxp\" (UID: \"0f71a336-51f3-4ad9-8dab-bc801ac5c0d9\") " pod="openstack/nova-cell1-83b2-account-create-5mvxp" Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.223727 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj55h\" (UniqueName: \"kubernetes.io/projected/0f71a336-51f3-4ad9-8dab-bc801ac5c0d9-kube-api-access-pj55h\") pod \"nova-cell1-83b2-account-create-5mvxp\" (UID: \"0f71a336-51f3-4ad9-8dab-bc801ac5c0d9\") " pod="openstack/nova-cell1-83b2-account-create-5mvxp" Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.396969 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-83b2-account-create-5mvxp" Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.432612 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0f81-account-create-brq6v"] Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.574839 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f522-account-create-rqblm"] Oct 05 08:52:02 crc kubenswrapper[4935]: W1005 08:52:02.577586 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca9810c5_05cd_440e_8b35_54940b7beb4f.slice/crio-b9e5905e1d8ad69c5254ed47014ce3a3d615c309cf0b8a86c99ec0c6df94fa8e WatchSource:0}: Error finding container b9e5905e1d8ad69c5254ed47014ce3a3d615c309cf0b8a86c99ec0c6df94fa8e: Status 404 returned error can't find the container with id b9e5905e1d8ad69c5254ed47014ce3a3d615c309cf0b8a86c99ec0c6df94fa8e Oct 05 08:52:02 crc kubenswrapper[4935]: W1005 08:52:02.628820 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f71a336_51f3_4ad9_8dab_bc801ac5c0d9.slice/crio-288b0e0fc2d34aa04cd66d08d4e5cb1ff95b960ed0843a71b3a788b1276b3a4b WatchSource:0}: Error finding container 288b0e0fc2d34aa04cd66d08d4e5cb1ff95b960ed0843a71b3a788b1276b3a4b: Status 404 returned error can't find the container with id 288b0e0fc2d34aa04cd66d08d4e5cb1ff95b960ed0843a71b3a788b1276b3a4b Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.631678 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-83b2-account-create-5mvxp"] Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.678597 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-83b2-account-create-5mvxp" event={"ID":"0f71a336-51f3-4ad9-8dab-bc801ac5c0d9","Type":"ContainerStarted","Data":"288b0e0fc2d34aa04cd66d08d4e5cb1ff95b960ed0843a71b3a788b1276b3a4b"} Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.679872 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f522-account-create-rqblm" event={"ID":"ca9810c5-05cd-440e-8b35-54940b7beb4f","Type":"ContainerStarted","Data":"b9e5905e1d8ad69c5254ed47014ce3a3d615c309cf0b8a86c99ec0c6df94fa8e"} Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.681624 4935 generic.go:334] "Generic (PLEG): container finished" podID="4c993f0b-5b16-4710-89ce-e5e1881479bb" containerID="01372cc20d63a99b396530eb9913597e8f692740c875d3e6ab7f8e2230e7568c" exitCode=0 Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.681718 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0f81-account-create-brq6v" event={"ID":"4c993f0b-5b16-4710-89ce-e5e1881479bb","Type":"ContainerDied","Data":"01372cc20d63a99b396530eb9913597e8f692740c875d3e6ab7f8e2230e7568c"} Oct 05 08:52:02 crc kubenswrapper[4935]: I1005 08:52:02.681909 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0f81-account-create-brq6v" event={"ID":"4c993f0b-5b16-4710-89ce-e5e1881479bb","Type":"ContainerStarted","Data":"3a55a242e6fdc4b81c613ceb5cd56931135a9288b466d9a5077754c519a57a7b"} Oct 05 08:52:03 crc kubenswrapper[4935]: I1005 08:52:03.693093 4935 generic.go:334] "Generic (PLEG): container finished" podID="0f71a336-51f3-4ad9-8dab-bc801ac5c0d9" containerID="38febbbc3e3f399bccafc6b2ffb7ae636a7298b1a6d204b13dac4aa8b98357a3" exitCode=0 Oct 05 08:52:03 crc kubenswrapper[4935]: I1005 08:52:03.693178 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-83b2-account-create-5mvxp" event={"ID":"0f71a336-51f3-4ad9-8dab-bc801ac5c0d9","Type":"ContainerDied","Data":"38febbbc3e3f399bccafc6b2ffb7ae636a7298b1a6d204b13dac4aa8b98357a3"} Oct 05 08:52:03 crc kubenswrapper[4935]: I1005 08:52:03.695429 4935 generic.go:334] "Generic (PLEG): container finished" podID="ca9810c5-05cd-440e-8b35-54940b7beb4f" containerID="e4d6661be0737c7116633e7c4d2f3eecc46901ad9577ad1a4a7f305acc479636" exitCode=0 Oct 05 08:52:03 crc kubenswrapper[4935]: I1005 08:52:03.695565 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f522-account-create-rqblm" event={"ID":"ca9810c5-05cd-440e-8b35-54940b7beb4f","Type":"ContainerDied","Data":"e4d6661be0737c7116633e7c4d2f3eecc46901ad9577ad1a4a7f305acc479636"} Oct 05 08:52:04 crc kubenswrapper[4935]: I1005 08:52:04.086421 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0f81-account-create-brq6v" Oct 05 08:52:04 crc kubenswrapper[4935]: I1005 08:52:04.243481 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmgm7\" (UniqueName: \"kubernetes.io/projected/4c993f0b-5b16-4710-89ce-e5e1881479bb-kube-api-access-kmgm7\") pod \"4c993f0b-5b16-4710-89ce-e5e1881479bb\" (UID: \"4c993f0b-5b16-4710-89ce-e5e1881479bb\") " Oct 05 08:52:04 crc kubenswrapper[4935]: I1005 08:52:04.251512 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c993f0b-5b16-4710-89ce-e5e1881479bb-kube-api-access-kmgm7" (OuterVolumeSpecName: "kube-api-access-kmgm7") pod "4c993f0b-5b16-4710-89ce-e5e1881479bb" (UID: "4c993f0b-5b16-4710-89ce-e5e1881479bb"). InnerVolumeSpecName "kube-api-access-kmgm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:04 crc kubenswrapper[4935]: I1005 08:52:04.345770 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmgm7\" (UniqueName: \"kubernetes.io/projected/4c993f0b-5b16-4710-89ce-e5e1881479bb-kube-api-access-kmgm7\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:04 crc kubenswrapper[4935]: I1005 08:52:04.706185 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0f81-account-create-brq6v" event={"ID":"4c993f0b-5b16-4710-89ce-e5e1881479bb","Type":"ContainerDied","Data":"3a55a242e6fdc4b81c613ceb5cd56931135a9288b466d9a5077754c519a57a7b"} Oct 05 08:52:04 crc kubenswrapper[4935]: I1005 08:52:04.706240 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a55a242e6fdc4b81c613ceb5cd56931135a9288b466d9a5077754c519a57a7b" Oct 05 08:52:04 crc kubenswrapper[4935]: I1005 08:52:04.706348 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0f81-account-create-brq6v" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.119205 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f522-account-create-rqblm" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.128807 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-83b2-account-create-5mvxp" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.264585 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj55h\" (UniqueName: \"kubernetes.io/projected/0f71a336-51f3-4ad9-8dab-bc801ac5c0d9-kube-api-access-pj55h\") pod \"0f71a336-51f3-4ad9-8dab-bc801ac5c0d9\" (UID: \"0f71a336-51f3-4ad9-8dab-bc801ac5c0d9\") " Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.264658 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh6bx\" (UniqueName: \"kubernetes.io/projected/ca9810c5-05cd-440e-8b35-54940b7beb4f-kube-api-access-wh6bx\") pod \"ca9810c5-05cd-440e-8b35-54940b7beb4f\" (UID: \"ca9810c5-05cd-440e-8b35-54940b7beb4f\") " Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.268742 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca9810c5-05cd-440e-8b35-54940b7beb4f-kube-api-access-wh6bx" (OuterVolumeSpecName: "kube-api-access-wh6bx") pod "ca9810c5-05cd-440e-8b35-54940b7beb4f" (UID: "ca9810c5-05cd-440e-8b35-54940b7beb4f"). InnerVolumeSpecName "kube-api-access-wh6bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.269644 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f71a336-51f3-4ad9-8dab-bc801ac5c0d9-kube-api-access-pj55h" (OuterVolumeSpecName: "kube-api-access-pj55h") pod "0f71a336-51f3-4ad9-8dab-bc801ac5c0d9" (UID: "0f71a336-51f3-4ad9-8dab-bc801ac5c0d9"). InnerVolumeSpecName "kube-api-access-pj55h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.366642 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh6bx\" (UniqueName: \"kubernetes.io/projected/ca9810c5-05cd-440e-8b35-54940b7beb4f-kube-api-access-wh6bx\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.366674 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj55h\" (UniqueName: \"kubernetes.io/projected/0f71a336-51f3-4ad9-8dab-bc801ac5c0d9-kube-api-access-pj55h\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.718578 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-83b2-account-create-5mvxp" event={"ID":"0f71a336-51f3-4ad9-8dab-bc801ac5c0d9","Type":"ContainerDied","Data":"288b0e0fc2d34aa04cd66d08d4e5cb1ff95b960ed0843a71b3a788b1276b3a4b"} Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.718622 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="288b0e0fc2d34aa04cd66d08d4e5cb1ff95b960ed0843a71b3a788b1276b3a4b" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.718618 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-83b2-account-create-5mvxp" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.721318 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f522-account-create-rqblm" event={"ID":"ca9810c5-05cd-440e-8b35-54940b7beb4f","Type":"ContainerDied","Data":"b9e5905e1d8ad69c5254ed47014ce3a3d615c309cf0b8a86c99ec0c6df94fa8e"} Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.721363 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9e5905e1d8ad69c5254ed47014ce3a3d615c309cf0b8a86c99ec0c6df94fa8e" Oct 05 08:52:05 crc kubenswrapper[4935]: I1005 08:52:05.721435 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f522-account-create-rqblm" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.914066 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kbvgl"] Oct 05 08:52:06 crc kubenswrapper[4935]: E1005 08:52:06.917798 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c993f0b-5b16-4710-89ce-e5e1881479bb" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.917834 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c993f0b-5b16-4710-89ce-e5e1881479bb" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: E1005 08:52:06.917870 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9810c5-05cd-440e-8b35-54940b7beb4f" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.917882 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9810c5-05cd-440e-8b35-54940b7beb4f" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: E1005 08:52:06.917919 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f71a336-51f3-4ad9-8dab-bc801ac5c0d9" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.917933 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f71a336-51f3-4ad9-8dab-bc801ac5c0d9" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.918284 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f71a336-51f3-4ad9-8dab-bc801ac5c0d9" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.918317 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c993f0b-5b16-4710-89ce-e5e1881479bb" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.918331 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9810c5-05cd-440e-8b35-54940b7beb4f" containerName="mariadb-account-create" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.919085 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.920966 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-q5df2" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.921349 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.921402 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 08:52:06 crc kubenswrapper[4935]: I1005 08:52:06.928674 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kbvgl"] Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.098878 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7ct4\" (UniqueName: \"kubernetes.io/projected/14ce69a4-c3f0-4585-893a-8da9ff928ce1-kube-api-access-r7ct4\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.098975 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.099002 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-config-data\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.099024 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-scripts\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.201219 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7ct4\" (UniqueName: \"kubernetes.io/projected/14ce69a4-c3f0-4585-893a-8da9ff928ce1-kube-api-access-r7ct4\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.201454 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.201611 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-config-data\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.201668 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-scripts\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.214873 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-scripts\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.215987 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.216879 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-config-data\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.223074 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7ct4\" (UniqueName: \"kubernetes.io/projected/14ce69a4-c3f0-4585-893a-8da9ff928ce1-kube-api-access-r7ct4\") pod \"nova-cell0-conductor-db-sync-kbvgl\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.248961 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.715079 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kbvgl"] Oct 05 08:52:07 crc kubenswrapper[4935]: I1005 08:52:07.744839 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" event={"ID":"14ce69a4-c3f0-4585-893a-8da9ff928ce1","Type":"ContainerStarted","Data":"6468937ede6b012f80a22cc122444b08cae25efd42027145cc7a9ac0c54faf37"} Oct 05 08:52:12 crc kubenswrapper[4935]: I1005 08:52:12.778776 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:52:12 crc kubenswrapper[4935]: E1005 08:52:12.780362 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:52:16 crc kubenswrapper[4935]: I1005 08:52:16.826613 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" event={"ID":"14ce69a4-c3f0-4585-893a-8da9ff928ce1","Type":"ContainerStarted","Data":"c6f2d5876ec0bd8c72fddf763c71a407f2d376afccab2664eed0b703fcc0e468"} Oct 05 08:52:16 crc kubenswrapper[4935]: I1005 08:52:16.852474 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" podStartSLOduration=2.4905416750000002 podStartE2EDuration="10.852449139s" podCreationTimestamp="2025-10-05 08:52:06 +0000 UTC" firstStartedPulling="2025-10-05 08:52:07.725690728 +0000 UTC m=+7161.608317188" lastFinishedPulling="2025-10-05 08:52:16.087598172 +0000 UTC m=+7169.970224652" observedRunningTime="2025-10-05 08:52:16.842250479 +0000 UTC m=+7170.724876969" watchObservedRunningTime="2025-10-05 08:52:16.852449139 +0000 UTC m=+7170.735075629" Oct 05 08:52:21 crc kubenswrapper[4935]: I1005 08:52:21.878470 4935 generic.go:334] "Generic (PLEG): container finished" podID="14ce69a4-c3f0-4585-893a-8da9ff928ce1" containerID="c6f2d5876ec0bd8c72fddf763c71a407f2d376afccab2664eed0b703fcc0e468" exitCode=0 Oct 05 08:52:21 crc kubenswrapper[4935]: I1005 08:52:21.878581 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" event={"ID":"14ce69a4-c3f0-4585-893a-8da9ff928ce1","Type":"ContainerDied","Data":"c6f2d5876ec0bd8c72fddf763c71a407f2d376afccab2664eed0b703fcc0e468"} Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.263599 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.444270 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7ct4\" (UniqueName: \"kubernetes.io/projected/14ce69a4-c3f0-4585-893a-8da9ff928ce1-kube-api-access-r7ct4\") pod \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.444365 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-combined-ca-bundle\") pod \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.444382 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-scripts\") pod \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.444519 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-config-data\") pod \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\" (UID: \"14ce69a4-c3f0-4585-893a-8da9ff928ce1\") " Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.450608 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ce69a4-c3f0-4585-893a-8da9ff928ce1-kube-api-access-r7ct4" (OuterVolumeSpecName: "kube-api-access-r7ct4") pod "14ce69a4-c3f0-4585-893a-8da9ff928ce1" (UID: "14ce69a4-c3f0-4585-893a-8da9ff928ce1"). InnerVolumeSpecName "kube-api-access-r7ct4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.451203 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-scripts" (OuterVolumeSpecName: "scripts") pod "14ce69a4-c3f0-4585-893a-8da9ff928ce1" (UID: "14ce69a4-c3f0-4585-893a-8da9ff928ce1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.470990 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-config-data" (OuterVolumeSpecName: "config-data") pod "14ce69a4-c3f0-4585-893a-8da9ff928ce1" (UID: "14ce69a4-c3f0-4585-893a-8da9ff928ce1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.475037 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14ce69a4-c3f0-4585-893a-8da9ff928ce1" (UID: "14ce69a4-c3f0-4585-893a-8da9ff928ce1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.546692 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7ct4\" (UniqueName: \"kubernetes.io/projected/14ce69a4-c3f0-4585-893a-8da9ff928ce1-kube-api-access-r7ct4\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.546726 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.546735 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.546743 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14ce69a4-c3f0-4585-893a-8da9ff928ce1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.777011 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:52:23 crc kubenswrapper[4935]: E1005 08:52:23.777503 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.910250 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" event={"ID":"14ce69a4-c3f0-4585-893a-8da9ff928ce1","Type":"ContainerDied","Data":"6468937ede6b012f80a22cc122444b08cae25efd42027145cc7a9ac0c54faf37"} Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.910316 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6468937ede6b012f80a22cc122444b08cae25efd42027145cc7a9ac0c54faf37" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.910316 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kbvgl" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.994052 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:52:23 crc kubenswrapper[4935]: E1005 08:52:23.994410 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ce69a4-c3f0-4585-893a-8da9ff928ce1" containerName="nova-cell0-conductor-db-sync" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.994424 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ce69a4-c3f0-4585-893a-8da9ff928ce1" containerName="nova-cell0-conductor-db-sync" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.994660 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ce69a4-c3f0-4585-893a-8da9ff928ce1" containerName="nova-cell0-conductor-db-sync" Oct 05 08:52:23 crc kubenswrapper[4935]: I1005 08:52:23.995278 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.001150 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.003149 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-q5df2" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.012757 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.159809 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqnp7\" (UniqueName: \"kubernetes.io/projected/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-kube-api-access-sqnp7\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.159907 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.159976 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.261688 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqnp7\" (UniqueName: \"kubernetes.io/projected/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-kube-api-access-sqnp7\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.261812 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.261904 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.268157 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.270065 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.286235 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqnp7\" (UniqueName: \"kubernetes.io/projected/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-kube-api-access-sqnp7\") pod \"nova-cell0-conductor-0\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.313685 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.799949 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:52:24 crc kubenswrapper[4935]: I1005 08:52:24.917770 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"76e7a7c0-e906-4e2d-95e1-0062edf3ff36","Type":"ContainerStarted","Data":"b49133c1bbce63709e7bafd77cd07148f9fa7d50aa3afce7e9411ef417cb5bc3"} Oct 05 08:52:25 crc kubenswrapper[4935]: I1005 08:52:25.933623 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"76e7a7c0-e906-4e2d-95e1-0062edf3ff36","Type":"ContainerStarted","Data":"af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1"} Oct 05 08:52:25 crc kubenswrapper[4935]: I1005 08:52:25.934233 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:25 crc kubenswrapper[4935]: I1005 08:52:25.966539 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.966516866 podStartE2EDuration="2.966516866s" podCreationTimestamp="2025-10-05 08:52:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:25.961450862 +0000 UTC m=+7179.844077352" watchObservedRunningTime="2025-10-05 08:52:25.966516866 +0000 UTC m=+7179.849143326" Oct 05 08:52:29 crc kubenswrapper[4935]: I1005 08:52:29.360203 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 05 08:52:29 crc kubenswrapper[4935]: I1005 08:52:29.893051 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-g9j8m"] Oct 05 08:52:29 crc kubenswrapper[4935]: I1005 08:52:29.894141 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:29 crc kubenswrapper[4935]: I1005 08:52:29.896572 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 05 08:52:29 crc kubenswrapper[4935]: I1005 08:52:29.896917 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 05 08:52:29 crc kubenswrapper[4935]: I1005 08:52:29.904873 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9j8m"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.026600 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.028149 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.030500 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.054125 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.068864 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-config-data\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.068972 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-scripts\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.069014 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s294w\" (UniqueName: \"kubernetes.io/projected/710e234a-a567-428d-855a-dc5c5a695860-kube-api-access-s294w\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.069070 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.105259 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.107220 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.111198 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.129350 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.139803 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.141079 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.144278 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.151386 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.170694 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsgrg\" (UniqueName: \"kubernetes.io/projected/966eaf41-48ff-4110-bfab-eb465a2f2b0d-kube-api-access-zsgrg\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.171151 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.172457 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.172699 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-config-data\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.176385 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-scripts\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.176878 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s294w\" (UniqueName: \"kubernetes.io/projected/710e234a-a567-428d-855a-dc5c5a695860-kube-api-access-s294w\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.176952 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-config-data\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.184604 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-scripts\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.195141 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.201670 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-config-data\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.220742 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84dc58f955-ljs2d"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.222305 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.228568 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s294w\" (UniqueName: \"kubernetes.io/projected/710e234a-a567-428d-855a-dc5c5a695860-kube-api-access-s294w\") pod \"nova-cell0-cell-mapping-g9j8m\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.251833 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84dc58f955-ljs2d"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.270521 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.274505 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.277833 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.277962 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-config-data\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.290985 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsgrg\" (UniqueName: \"kubernetes.io/projected/966eaf41-48ff-4110-bfab-eb465a2f2b0d-kube-api-access-zsgrg\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.291083 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.291161 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-logs\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.291244 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.291299 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.291450 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmtzj\" (UniqueName: \"kubernetes.io/projected/0cd2e2a0-6247-45b3-a8c3-df740829908f-kube-api-access-qmtzj\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.291565 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.291618 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-config-data\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.291664 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw998\" (UniqueName: \"kubernetes.io/projected/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-kube-api-access-lw998\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.310534 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.315386 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-config-data\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.318704 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.332525 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsgrg\" (UniqueName: \"kubernetes.io/projected/966eaf41-48ff-4110-bfab-eb465a2f2b0d-kube-api-access-zsgrg\") pod \"nova-scheduler-0\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.360921 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.406755 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce78cb3-073c-430a-8c06-a584144cfcd1-logs\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.406840 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.406932 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-dns-svc\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.406954 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-config-data\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.407031 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw998\" (UniqueName: \"kubernetes.io/projected/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-kube-api-access-lw998\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.407370 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk9hj\" (UniqueName: \"kubernetes.io/projected/0ce78cb3-073c-430a-8c06-a584144cfcd1-kube-api-access-sk9hj\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.407405 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-nb\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409115 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409279 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409332 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-config-data\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409375 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-logs\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409435 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-config\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409473 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-sb\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409518 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409607 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf8rx\" (UniqueName: \"kubernetes.io/projected/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-kube-api-access-pf8rx\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.409659 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmtzj\" (UniqueName: \"kubernetes.io/projected/0cd2e2a0-6247-45b3-a8c3-df740829908f-kube-api-access-qmtzj\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.410087 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-logs\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.411716 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-config-data\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.413460 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.413962 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.416138 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.428842 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw998\" (UniqueName: \"kubernetes.io/projected/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-kube-api-access-lw998\") pod \"nova-metadata-0\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.439691 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmtzj\" (UniqueName: \"kubernetes.io/projected/0cd2e2a0-6247-45b3-a8c3-df740829908f-kube-api-access-qmtzj\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.444710 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.473410 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.514480 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515386 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515488 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-config-data\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515545 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-config\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515574 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-sb\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515639 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf8rx\" (UniqueName: \"kubernetes.io/projected/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-kube-api-access-pf8rx\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515702 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce78cb3-073c-430a-8c06-a584144cfcd1-logs\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-dns-svc\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515811 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk9hj\" (UniqueName: \"kubernetes.io/projected/0ce78cb3-073c-430a-8c06-a584144cfcd1-kube-api-access-sk9hj\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.515843 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-nb\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.516831 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-nb\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.518278 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce78cb3-073c-430a-8c06-a584144cfcd1-logs\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.520591 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-sb\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.521520 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-config-data\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.522715 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.522859 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-config\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.523029 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-dns-svc\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.539761 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf8rx\" (UniqueName: \"kubernetes.io/projected/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-kube-api-access-pf8rx\") pod \"dnsmasq-dns-84dc58f955-ljs2d\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.547654 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk9hj\" (UniqueName: \"kubernetes.io/projected/0ce78cb3-073c-430a-8c06-a584144cfcd1-kube-api-access-sk9hj\") pod \"nova-api-0\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.575203 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.731418 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:52:30 crc kubenswrapper[4935]: I1005 08:52:30.888255 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:30 crc kubenswrapper[4935]: W1005 08:52:30.901344 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod966eaf41_48ff_4110_bfab_eb465a2f2b0d.slice/crio-86ebfe7621813106a1c14dbaa2e7f6ee9b0215e156a7f7750335b78564244abc WatchSource:0}: Error finding container 86ebfe7621813106a1c14dbaa2e7f6ee9b0215e156a7f7750335b78564244abc: Status 404 returned error can't find the container with id 86ebfe7621813106a1c14dbaa2e7f6ee9b0215e156a7f7750335b78564244abc Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.000765 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"966eaf41-48ff-4110-bfab-eb465a2f2b0d","Type":"ContainerStarted","Data":"86ebfe7621813106a1c14dbaa2e7f6ee9b0215e156a7f7750335b78564244abc"} Oct 05 08:52:31 crc kubenswrapper[4935]: W1005 08:52:31.078356 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod710e234a_a567_428d_855a_dc5c5a695860.slice/crio-e731faad79dc57de6d88759c6e9bfafaed3c19d5238e36ee3a32e6ec6ae548d5 WatchSource:0}: Error finding container e731faad79dc57de6d88759c6e9bfafaed3c19d5238e36ee3a32e6ec6ae548d5: Status 404 returned error can't find the container with id e731faad79dc57de6d88759c6e9bfafaed3c19d5238e36ee3a32e6ec6ae548d5 Oct 05 08:52:31 crc kubenswrapper[4935]: W1005 08:52:31.087146 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cd2e2a0_6247_45b3_a8c3_df740829908f.slice/crio-959b9143bcd44010cca5b076ea126afd4305aecf3882f706b10fbbcc88bc19a2 WatchSource:0}: Error finding container 959b9143bcd44010cca5b076ea126afd4305aecf3882f706b10fbbcc88bc19a2: Status 404 returned error can't find the container with id 959b9143bcd44010cca5b076ea126afd4305aecf3882f706b10fbbcc88bc19a2 Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.090539 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:52:31 crc kubenswrapper[4935]: W1005 08:52:31.091875 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9025cbb2_c30d_450b_bd98_ad68dd90d9e5.slice/crio-2cee29443e86b1bcc4e1aad603266d4b4f7addfafcde9f819c6b692c79ea0a52 WatchSource:0}: Error finding container 2cee29443e86b1bcc4e1aad603266d4b4f7addfafcde9f819c6b692c79ea0a52: Status 404 returned error can't find the container with id 2cee29443e86b1bcc4e1aad603266d4b4f7addfafcde9f819c6b692c79ea0a52 Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.114357 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9j8m"] Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.131137 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.160824 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z52d2"] Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.162807 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.168900 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.169256 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.174106 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z52d2"] Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.215632 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84dc58f955-ljs2d"] Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.287980 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.340294 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.340419 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-scripts\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.340549 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-config-data\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.340736 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4mnq\" (UniqueName: \"kubernetes.io/projected/93d805da-2c17-4b2a-ad2f-609a03683de9-kube-api-access-b4mnq\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.442554 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.442597 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-scripts\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.442639 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-config-data\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.442708 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4mnq\" (UniqueName: \"kubernetes.io/projected/93d805da-2c17-4b2a-ad2f-609a03683de9-kube-api-access-b4mnq\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.453189 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-scripts\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.453329 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.453591 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-config-data\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.459554 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4mnq\" (UniqueName: \"kubernetes.io/projected/93d805da-2c17-4b2a-ad2f-609a03683de9-kube-api-access-b4mnq\") pod \"nova-cell1-conductor-db-sync-z52d2\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:31 crc kubenswrapper[4935]: I1005 08:52:31.482858 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.021767 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ce78cb3-073c-430a-8c06-a584144cfcd1","Type":"ContainerStarted","Data":"5997b1e64e456bfdc99daee7877e8f39b7f6eb24ebb828e57ef0bbdfad227ec2"} Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.027245 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z52d2"] Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.054883 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9025cbb2-c30d-450b-bd98-ad68dd90d9e5","Type":"ContainerStarted","Data":"2cee29443e86b1bcc4e1aad603266d4b4f7addfafcde9f819c6b692c79ea0a52"} Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.058013 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9j8m" event={"ID":"710e234a-a567-428d-855a-dc5c5a695860","Type":"ContainerStarted","Data":"b6fe144670d2f5db06ced874e32576eeaa621ab9f5420f3049766466efe7da56"} Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.058074 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9j8m" event={"ID":"710e234a-a567-428d-855a-dc5c5a695860","Type":"ContainerStarted","Data":"e731faad79dc57de6d88759c6e9bfafaed3c19d5238e36ee3a32e6ec6ae548d5"} Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.059984 4935 generic.go:334] "Generic (PLEG): container finished" podID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerID="8523e9c5ce4cde1a4397f2e533b4c940d0d1709fad8441355e1d552c5fc1aa7f" exitCode=0 Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.060045 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" event={"ID":"eb2cfe55-91a7-4893-839f-64b3d6d46c9f","Type":"ContainerDied","Data":"8523e9c5ce4cde1a4397f2e533b4c940d0d1709fad8441355e1d552c5fc1aa7f"} Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.060080 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" event={"ID":"eb2cfe55-91a7-4893-839f-64b3d6d46c9f","Type":"ContainerStarted","Data":"346548fa1c79c59699de1fb827fd30f47a0d739ea16ab3521bffb799513e5c67"} Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.061037 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0cd2e2a0-6247-45b3-a8c3-df740829908f","Type":"ContainerStarted","Data":"959b9143bcd44010cca5b076ea126afd4305aecf3882f706b10fbbcc88bc19a2"} Oct 05 08:52:32 crc kubenswrapper[4935]: I1005 08:52:32.078167 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-g9j8m" podStartSLOduration=3.078072979 podStartE2EDuration="3.078072979s" podCreationTimestamp="2025-10-05 08:52:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:32.074815233 +0000 UTC m=+7185.957441693" watchObservedRunningTime="2025-10-05 08:52:32.078072979 +0000 UTC m=+7185.960699439" Oct 05 08:52:32 crc kubenswrapper[4935]: W1005 08:52:32.864281 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93d805da_2c17_4b2a_ad2f_609a03683de9.slice/crio-388bd5fcde85a1b32a976b6ac5f2b01c5442b159ca44bdcd37eacc5ea26813ae WatchSource:0}: Error finding container 388bd5fcde85a1b32a976b6ac5f2b01c5442b159ca44bdcd37eacc5ea26813ae: Status 404 returned error can't find the container with id 388bd5fcde85a1b32a976b6ac5f2b01c5442b159ca44bdcd37eacc5ea26813ae Oct 05 08:52:33 crc kubenswrapper[4935]: I1005 08:52:33.073058 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z52d2" event={"ID":"93d805da-2c17-4b2a-ad2f-609a03683de9","Type":"ContainerStarted","Data":"388bd5fcde85a1b32a976b6ac5f2b01c5442b159ca44bdcd37eacc5ea26813ae"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.082662 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ce78cb3-073c-430a-8c06-a584144cfcd1","Type":"ContainerStarted","Data":"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.083070 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ce78cb3-073c-430a-8c06-a584144cfcd1","Type":"ContainerStarted","Data":"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.086104 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"966eaf41-48ff-4110-bfab-eb465a2f2b0d","Type":"ContainerStarted","Data":"9284b372033818c061ffc34b2a5d36731698ac3d2fa520e0cafa6b3f5d158352"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.091515 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9025cbb2-c30d-450b-bd98-ad68dd90d9e5","Type":"ContainerStarted","Data":"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.091561 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9025cbb2-c30d-450b-bd98-ad68dd90d9e5","Type":"ContainerStarted","Data":"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.093425 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z52d2" event={"ID":"93d805da-2c17-4b2a-ad2f-609a03683de9","Type":"ContainerStarted","Data":"e36244e8ba12537dfd22c4da79722910b3092fe0975fdf32c75ed4b144f4b0fa"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.095832 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" event={"ID":"eb2cfe55-91a7-4893-839f-64b3d6d46c9f","Type":"ContainerStarted","Data":"1dc48e927d06f30f333ea6f79b420948c1d413f840ac9cfa35e9cb1568cc09f7"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.096004 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.097200 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0cd2e2a0-6247-45b3-a8c3-df740829908f","Type":"ContainerStarted","Data":"4879cd0b51ae4570f9468ff413c3752a280ad7596a1ad09ecfa714142128d33e"} Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.108352 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.129737119 podStartE2EDuration="4.108337293s" podCreationTimestamp="2025-10-05 08:52:30 +0000 UTC" firstStartedPulling="2025-10-05 08:52:31.304673964 +0000 UTC m=+7185.187300434" lastFinishedPulling="2025-10-05 08:52:33.283274148 +0000 UTC m=+7187.165900608" observedRunningTime="2025-10-05 08:52:34.102905689 +0000 UTC m=+7187.985532149" watchObservedRunningTime="2025-10-05 08:52:34.108337293 +0000 UTC m=+7187.990963753" Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.126671 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.9354600450000001 podStartE2EDuration="4.126655109s" podCreationTimestamp="2025-10-05 08:52:30 +0000 UTC" firstStartedPulling="2025-10-05 08:52:31.091117989 +0000 UTC m=+7184.973744449" lastFinishedPulling="2025-10-05 08:52:33.282313043 +0000 UTC m=+7187.164939513" observedRunningTime="2025-10-05 08:52:34.125128438 +0000 UTC m=+7188.007754898" watchObservedRunningTime="2025-10-05 08:52:34.126655109 +0000 UTC m=+7188.009281569" Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.149229 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-z52d2" podStartSLOduration=3.149211037 podStartE2EDuration="3.149211037s" podCreationTimestamp="2025-10-05 08:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:34.144223505 +0000 UTC m=+7188.026849965" watchObservedRunningTime="2025-10-05 08:52:34.149211037 +0000 UTC m=+7188.031837497" Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.170795 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" podStartSLOduration=4.170768249 podStartE2EDuration="4.170768249s" podCreationTimestamp="2025-10-05 08:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:34.160494076 +0000 UTC m=+7188.043120536" watchObservedRunningTime="2025-10-05 08:52:34.170768249 +0000 UTC m=+7188.053394709" Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.180905 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.808571309 podStartE2EDuration="5.180875087s" podCreationTimestamp="2025-10-05 08:52:29 +0000 UTC" firstStartedPulling="2025-10-05 08:52:30.906983295 +0000 UTC m=+7184.789609755" lastFinishedPulling="2025-10-05 08:52:33.279287053 +0000 UTC m=+7187.161913533" observedRunningTime="2025-10-05 08:52:34.173370208 +0000 UTC m=+7188.055996678" watchObservedRunningTime="2025-10-05 08:52:34.180875087 +0000 UTC m=+7188.063501547" Oct 05 08:52:34 crc kubenswrapper[4935]: I1005 08:52:34.196281 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.013038613 podStartE2EDuration="4.196259035s" podCreationTimestamp="2025-10-05 08:52:30 +0000 UTC" firstStartedPulling="2025-10-05 08:52:31.099865061 +0000 UTC m=+7184.982491521" lastFinishedPulling="2025-10-05 08:52:33.283085483 +0000 UTC m=+7187.165711943" observedRunningTime="2025-10-05 08:52:34.190822621 +0000 UTC m=+7188.073449091" watchObservedRunningTime="2025-10-05 08:52:34.196259035 +0000 UTC m=+7188.078885495" Oct 05 08:52:35 crc kubenswrapper[4935]: I1005 08:52:35.361923 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 08:52:35 crc kubenswrapper[4935]: I1005 08:52:35.446760 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:52:35 crc kubenswrapper[4935]: I1005 08:52:35.446811 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:52:35 crc kubenswrapper[4935]: I1005 08:52:35.475144 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:36 crc kubenswrapper[4935]: I1005 08:52:36.118942 4935 generic.go:334] "Generic (PLEG): container finished" podID="710e234a-a567-428d-855a-dc5c5a695860" containerID="b6fe144670d2f5db06ced874e32576eeaa621ab9f5420f3049766466efe7da56" exitCode=0 Oct 05 08:52:36 crc kubenswrapper[4935]: I1005 08:52:36.119363 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9j8m" event={"ID":"710e234a-a567-428d-855a-dc5c5a695860","Type":"ContainerDied","Data":"b6fe144670d2f5db06ced874e32576eeaa621ab9f5420f3049766466efe7da56"} Oct 05 08:52:36 crc kubenswrapper[4935]: I1005 08:52:36.805712 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:52:36 crc kubenswrapper[4935]: E1005 08:52:36.806327 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.129440 4935 generic.go:334] "Generic (PLEG): container finished" podID="93d805da-2c17-4b2a-ad2f-609a03683de9" containerID="e36244e8ba12537dfd22c4da79722910b3092fe0975fdf32c75ed4b144f4b0fa" exitCode=0 Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.129551 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z52d2" event={"ID":"93d805da-2c17-4b2a-ad2f-609a03683de9","Type":"ContainerDied","Data":"e36244e8ba12537dfd22c4da79722910b3092fe0975fdf32c75ed4b144f4b0fa"} Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.536415 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.677753 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-scripts\") pod \"710e234a-a567-428d-855a-dc5c5a695860\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.677944 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s294w\" (UniqueName: \"kubernetes.io/projected/710e234a-a567-428d-855a-dc5c5a695860-kube-api-access-s294w\") pod \"710e234a-a567-428d-855a-dc5c5a695860\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.678017 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-combined-ca-bundle\") pod \"710e234a-a567-428d-855a-dc5c5a695860\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.678090 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-config-data\") pod \"710e234a-a567-428d-855a-dc5c5a695860\" (UID: \"710e234a-a567-428d-855a-dc5c5a695860\") " Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.684599 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-scripts" (OuterVolumeSpecName: "scripts") pod "710e234a-a567-428d-855a-dc5c5a695860" (UID: "710e234a-a567-428d-855a-dc5c5a695860"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.685268 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/710e234a-a567-428d-855a-dc5c5a695860-kube-api-access-s294w" (OuterVolumeSpecName: "kube-api-access-s294w") pod "710e234a-a567-428d-855a-dc5c5a695860" (UID: "710e234a-a567-428d-855a-dc5c5a695860"). InnerVolumeSpecName "kube-api-access-s294w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.723598 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-config-data" (OuterVolumeSpecName: "config-data") pod "710e234a-a567-428d-855a-dc5c5a695860" (UID: "710e234a-a567-428d-855a-dc5c5a695860"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.726866 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "710e234a-a567-428d-855a-dc5c5a695860" (UID: "710e234a-a567-428d-855a-dc5c5a695860"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.780275 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.780321 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.780338 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s294w\" (UniqueName: \"kubernetes.io/projected/710e234a-a567-428d-855a-dc5c5a695860-kube-api-access-s294w\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:37 crc kubenswrapper[4935]: I1005 08:52:37.780352 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710e234a-a567-428d-855a-dc5c5a695860-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.143576 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9j8m" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.143570 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9j8m" event={"ID":"710e234a-a567-428d-855a-dc5c5a695860","Type":"ContainerDied","Data":"e731faad79dc57de6d88759c6e9bfafaed3c19d5238e36ee3a32e6ec6ae548d5"} Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.145771 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e731faad79dc57de6d88759c6e9bfafaed3c19d5238e36ee3a32e6ec6ae548d5" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.255294 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.255503 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerName="nova-api-log" containerID="cri-o://905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07" gracePeriod=30 Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.255923 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerName="nova-api-api" containerID="cri-o://b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d" gracePeriod=30 Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.278411 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.278613 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="966eaf41-48ff-4110-bfab-eb465a2f2b0d" containerName="nova-scheduler-scheduler" containerID="cri-o://9284b372033818c061ffc34b2a5d36731698ac3d2fa520e0cafa6b3f5d158352" gracePeriod=30 Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.317616 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.317822 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerName="nova-metadata-log" containerID="cri-o://0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766" gracePeriod=30 Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.317974 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerName="nova-metadata-metadata" containerID="cri-o://1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22" gracePeriod=30 Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.706151 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.804591 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-scripts\") pod \"93d805da-2c17-4b2a-ad2f-609a03683de9\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.806430 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-config-data\") pod \"93d805da-2c17-4b2a-ad2f-609a03683de9\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.806724 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4mnq\" (UniqueName: \"kubernetes.io/projected/93d805da-2c17-4b2a-ad2f-609a03683de9-kube-api-access-b4mnq\") pod \"93d805da-2c17-4b2a-ad2f-609a03683de9\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.806789 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-combined-ca-bundle\") pod \"93d805da-2c17-4b2a-ad2f-609a03683de9\" (UID: \"93d805da-2c17-4b2a-ad2f-609a03683de9\") " Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.817134 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-scripts" (OuterVolumeSpecName: "scripts") pod "93d805da-2c17-4b2a-ad2f-609a03683de9" (UID: "93d805da-2c17-4b2a-ad2f-609a03683de9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.824921 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93d805da-2c17-4b2a-ad2f-609a03683de9-kube-api-access-b4mnq" (OuterVolumeSpecName: "kube-api-access-b4mnq") pod "93d805da-2c17-4b2a-ad2f-609a03683de9" (UID: "93d805da-2c17-4b2a-ad2f-609a03683de9"). InnerVolumeSpecName "kube-api-access-b4mnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.838368 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93d805da-2c17-4b2a-ad2f-609a03683de9" (UID: "93d805da-2c17-4b2a-ad2f-609a03683de9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.869949 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-config-data" (OuterVolumeSpecName: "config-data") pod "93d805da-2c17-4b2a-ad2f-609a03683de9" (UID: "93d805da-2c17-4b2a-ad2f-609a03683de9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.908851 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4mnq\" (UniqueName: \"kubernetes.io/projected/93d805da-2c17-4b2a-ad2f-609a03683de9-kube-api-access-b4mnq\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.908884 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.909186 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.909201 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d805da-2c17-4b2a-ad2f-609a03683de9-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.958716 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:52:38 crc kubenswrapper[4935]: I1005 08:52:38.966390 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112029 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-logs\") pod \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112112 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk9hj\" (UniqueName: \"kubernetes.io/projected/0ce78cb3-073c-430a-8c06-a584144cfcd1-kube-api-access-sk9hj\") pod \"0ce78cb3-073c-430a-8c06-a584144cfcd1\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112180 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-config-data\") pod \"0ce78cb3-073c-430a-8c06-a584144cfcd1\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112232 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce78cb3-073c-430a-8c06-a584144cfcd1-logs\") pod \"0ce78cb3-073c-430a-8c06-a584144cfcd1\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112296 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-combined-ca-bundle\") pod \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112337 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-config-data\") pod \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112387 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-combined-ca-bundle\") pod \"0ce78cb3-073c-430a-8c06-a584144cfcd1\" (UID: \"0ce78cb3-073c-430a-8c06-a584144cfcd1\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112460 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw998\" (UniqueName: \"kubernetes.io/projected/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-kube-api-access-lw998\") pod \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\" (UID: \"9025cbb2-c30d-450b-bd98-ad68dd90d9e5\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112734 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-logs" (OuterVolumeSpecName: "logs") pod "9025cbb2-c30d-450b-bd98-ad68dd90d9e5" (UID: "9025cbb2-c30d-450b-bd98-ad68dd90d9e5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.112832 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ce78cb3-073c-430a-8c06-a584144cfcd1-logs" (OuterVolumeSpecName: "logs") pod "0ce78cb3-073c-430a-8c06-a584144cfcd1" (UID: "0ce78cb3-073c-430a-8c06-a584144cfcd1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.113588 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce78cb3-073c-430a-8c06-a584144cfcd1-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.113619 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.138646 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-kube-api-access-lw998" (OuterVolumeSpecName: "kube-api-access-lw998") pod "9025cbb2-c30d-450b-bd98-ad68dd90d9e5" (UID: "9025cbb2-c30d-450b-bd98-ad68dd90d9e5"). InnerVolumeSpecName "kube-api-access-lw998". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.139218 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ce78cb3-073c-430a-8c06-a584144cfcd1-kube-api-access-sk9hj" (OuterVolumeSpecName: "kube-api-access-sk9hj") pod "0ce78cb3-073c-430a-8c06-a584144cfcd1" (UID: "0ce78cb3-073c-430a-8c06-a584144cfcd1"). InnerVolumeSpecName "kube-api-access-sk9hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.143424 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-config-data" (OuterVolumeSpecName: "config-data") pod "0ce78cb3-073c-430a-8c06-a584144cfcd1" (UID: "0ce78cb3-073c-430a-8c06-a584144cfcd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.143594 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ce78cb3-073c-430a-8c06-a584144cfcd1" (UID: "0ce78cb3-073c-430a-8c06-a584144cfcd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.151927 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9025cbb2-c30d-450b-bd98-ad68dd90d9e5" (UID: "9025cbb2-c30d-450b-bd98-ad68dd90d9e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.151949 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-config-data" (OuterVolumeSpecName: "config-data") pod "9025cbb2-c30d-450b-bd98-ad68dd90d9e5" (UID: "9025cbb2-c30d-450b-bd98-ad68dd90d9e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.155454 4935 generic.go:334] "Generic (PLEG): container finished" podID="966eaf41-48ff-4110-bfab-eb465a2f2b0d" containerID="9284b372033818c061ffc34b2a5d36731698ac3d2fa520e0cafa6b3f5d158352" exitCode=0 Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.155751 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"966eaf41-48ff-4110-bfab-eb465a2f2b0d","Type":"ContainerDied","Data":"9284b372033818c061ffc34b2a5d36731698ac3d2fa520e0cafa6b3f5d158352"} Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.157757 4935 generic.go:334] "Generic (PLEG): container finished" podID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerID="1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22" exitCode=0 Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.157779 4935 generic.go:334] "Generic (PLEG): container finished" podID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerID="0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766" exitCode=143 Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.157843 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.157870 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9025cbb2-c30d-450b-bd98-ad68dd90d9e5","Type":"ContainerDied","Data":"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22"} Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.157941 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9025cbb2-c30d-450b-bd98-ad68dd90d9e5","Type":"ContainerDied","Data":"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766"} Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.157955 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9025cbb2-c30d-450b-bd98-ad68dd90d9e5","Type":"ContainerDied","Data":"2cee29443e86b1bcc4e1aad603266d4b4f7addfafcde9f819c6b692c79ea0a52"} Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.157980 4935 scope.go:117] "RemoveContainer" containerID="1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.168502 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z52d2" event={"ID":"93d805da-2c17-4b2a-ad2f-609a03683de9","Type":"ContainerDied","Data":"388bd5fcde85a1b32a976b6ac5f2b01c5442b159ca44bdcd37eacc5ea26813ae"} Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.168534 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z52d2" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.168551 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="388bd5fcde85a1b32a976b6ac5f2b01c5442b159ca44bdcd37eacc5ea26813ae" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.183874 4935 generic.go:334] "Generic (PLEG): container finished" podID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerID="b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d" exitCode=0 Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.183924 4935 generic.go:334] "Generic (PLEG): container finished" podID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerID="905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07" exitCode=143 Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.183957 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ce78cb3-073c-430a-8c06-a584144cfcd1","Type":"ContainerDied","Data":"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d"} Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.183997 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ce78cb3-073c-430a-8c06-a584144cfcd1","Type":"ContainerDied","Data":"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07"} Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.184013 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ce78cb3-073c-430a-8c06-a584144cfcd1","Type":"ContainerDied","Data":"5997b1e64e456bfdc99daee7877e8f39b7f6eb24ebb828e57ef0bbdfad227ec2"} Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.184110 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.259037 4935 scope.go:117] "RemoveContainer" containerID="0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.279987 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.287079 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk9hj\" (UniqueName: \"kubernetes.io/projected/0ce78cb3-073c-430a-8c06-a584144cfcd1-kube-api-access-sk9hj\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.287121 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.287131 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.287145 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.287154 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce78cb3-073c-430a-8c06-a584144cfcd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.287163 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw998\" (UniqueName: \"kubernetes.io/projected/9025cbb2-c30d-450b-bd98-ad68dd90d9e5-kube-api-access-lw998\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.300149 4935 scope.go:117] "RemoveContainer" containerID="1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.301956 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22\": container with ID starting with 1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22 not found: ID does not exist" containerID="1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.302022 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22"} err="failed to get container status \"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22\": rpc error: code = NotFound desc = could not find container \"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22\": container with ID starting with 1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22 not found: ID does not exist" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.302058 4935 scope.go:117] "RemoveContainer" containerID="0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.302651 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766\": container with ID starting with 0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766 not found: ID does not exist" containerID="0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.302689 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766"} err="failed to get container status \"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766\": rpc error: code = NotFound desc = could not find container \"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766\": container with ID starting with 0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766 not found: ID does not exist" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.302713 4935 scope.go:117] "RemoveContainer" containerID="1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.303175 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22"} err="failed to get container status \"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22\": rpc error: code = NotFound desc = could not find container \"1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22\": container with ID starting with 1a0297b68365dde4be5322549adb96bed3d160b2ffbef7f5305c89ae10823e22 not found: ID does not exist" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.303205 4935 scope.go:117] "RemoveContainer" containerID="0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.304645 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766"} err="failed to get container status \"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766\": rpc error: code = NotFound desc = could not find container \"0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766\": container with ID starting with 0ded574c05aafc4601e7ad934d47c8653da5230907dde2fb197b7ee62f37a766 not found: ID does not exist" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.304771 4935 scope.go:117] "RemoveContainer" containerID="b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.319674 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.333594 4935 scope.go:117] "RemoveContainer" containerID="905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.343966 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.344373 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerName="nova-metadata-log" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344467 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerName="nova-metadata-log" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.344483 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d805da-2c17-4b2a-ad2f-609a03683de9" containerName="nova-cell1-conductor-db-sync" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344489 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d805da-2c17-4b2a-ad2f-609a03683de9" containerName="nova-cell1-conductor-db-sync" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.344498 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerName="nova-api-api" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344518 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerName="nova-api-api" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.344528 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerName="nova-api-log" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344534 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerName="nova-api-log" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.344557 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerName="nova-metadata-metadata" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344619 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerName="nova-metadata-metadata" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.344645 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="710e234a-a567-428d-855a-dc5c5a695860" containerName="nova-manage" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344651 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="710e234a-a567-428d-855a-dc5c5a695860" containerName="nova-manage" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344877 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerName="nova-metadata-metadata" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344929 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerName="nova-api-api" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344941 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" containerName="nova-api-log" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344955 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="93d805da-2c17-4b2a-ad2f-609a03683de9" containerName="nova-cell1-conductor-db-sync" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344971 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" containerName="nova-metadata-log" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.344980 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="710e234a-a567-428d-855a-dc5c5a695860" containerName="nova-manage" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.345965 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.347145 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.353648 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.368418 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.373272 4935 scope.go:117] "RemoveContainer" containerID="b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.374500 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d\": container with ID starting with b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d not found: ID does not exist" containerID="b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.374536 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d"} err="failed to get container status \"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d\": rpc error: code = NotFound desc = could not find container \"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d\": container with ID starting with b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d not found: ID does not exist" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.374560 4935 scope.go:117] "RemoveContainer" containerID="905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07" Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.376547 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07\": container with ID starting with 905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07 not found: ID does not exist" containerID="905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.376590 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07"} err="failed to get container status \"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07\": rpc error: code = NotFound desc = could not find container \"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07\": container with ID starting with 905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07 not found: ID does not exist" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.376610 4935 scope.go:117] "RemoveContainer" containerID="b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.377792 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d"} err="failed to get container status \"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d\": rpc error: code = NotFound desc = could not find container \"b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d\": container with ID starting with b32776d902d43bf9a11c30d9e7ab3aaaa19c15696cd645b88209e95e50d33e2d not found: ID does not exist" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.377848 4935 scope.go:117] "RemoveContainer" containerID="905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.378477 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07"} err="failed to get container status \"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07\": rpc error: code = NotFound desc = could not find container \"905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07\": container with ID starting with 905e9313f839cced252015aab4ec2e47c2d9ab0fc93e2af6b0de01f6c7696c07 not found: ID does not exist" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.380760 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: E1005 08:52:39.381209 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="966eaf41-48ff-4110-bfab-eb465a2f2b0d" containerName="nova-scheduler-scheduler" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.381224 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="966eaf41-48ff-4110-bfab-eb465a2f2b0d" containerName="nova-scheduler-scheduler" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.381434 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="966eaf41-48ff-4110-bfab-eb465a2f2b0d" containerName="nova-scheduler-scheduler" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.382089 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.391940 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-config-data\") pod \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.391981 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-combined-ca-bundle\") pod \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.392050 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsgrg\" (UniqueName: \"kubernetes.io/projected/966eaf41-48ff-4110-bfab-eb465a2f2b0d-kube-api-access-zsgrg\") pod \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\" (UID: \"966eaf41-48ff-4110-bfab-eb465a2f2b0d\") " Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.392532 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tphz2\" (UniqueName: \"kubernetes.io/projected/4c088042-dec3-4afd-918b-f0ca1615e88d-kube-api-access-tphz2\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.392598 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c088042-dec3-4afd-918b-f0ca1615e88d-logs\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.392619 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.392641 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-config-data\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.396757 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.398812 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/966eaf41-48ff-4110-bfab-eb465a2f2b0d-kube-api-access-zsgrg" (OuterVolumeSpecName: "kube-api-access-zsgrg") pod "966eaf41-48ff-4110-bfab-eb465a2f2b0d" (UID: "966eaf41-48ff-4110-bfab-eb465a2f2b0d"). InnerVolumeSpecName "kube-api-access-zsgrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.414672 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.417449 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-config-data" (OuterVolumeSpecName: "config-data") pod "966eaf41-48ff-4110-bfab-eb465a2f2b0d" (UID: "966eaf41-48ff-4110-bfab-eb465a2f2b0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.422039 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.426224 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "966eaf41-48ff-4110-bfab-eb465a2f2b0d" (UID: "966eaf41-48ff-4110-bfab-eb465a2f2b0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.428189 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.435740 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.437237 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.439475 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.442653 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.494908 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/230ff47c-ce58-4669-ba80-4a019365919d-logs\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495009 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-config-data\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495045 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d76v6\" (UniqueName: \"kubernetes.io/projected/2d0f44d4-a284-4c21-bea7-b964477081c4-kube-api-access-d76v6\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495261 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495344 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tphz2\" (UniqueName: \"kubernetes.io/projected/4c088042-dec3-4afd-918b-f0ca1615e88d-kube-api-access-tphz2\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495510 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr2zd\" (UniqueName: \"kubernetes.io/projected/230ff47c-ce58-4669-ba80-4a019365919d-kube-api-access-fr2zd\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495594 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c088042-dec3-4afd-918b-f0ca1615e88d-logs\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495646 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495664 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495689 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495736 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-config-data\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495869 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495881 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/966eaf41-48ff-4110-bfab-eb465a2f2b0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.495910 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsgrg\" (UniqueName: \"kubernetes.io/projected/966eaf41-48ff-4110-bfab-eb465a2f2b0d-kube-api-access-zsgrg\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.496876 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c088042-dec3-4afd-918b-f0ca1615e88d-logs\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.500219 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-config-data\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.500737 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.511332 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tphz2\" (UniqueName: \"kubernetes.io/projected/4c088042-dec3-4afd-918b-f0ca1615e88d-kube-api-access-tphz2\") pod \"nova-metadata-0\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.597882 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr2zd\" (UniqueName: \"kubernetes.io/projected/230ff47c-ce58-4669-ba80-4a019365919d-kube-api-access-fr2zd\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.597965 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.597986 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.598020 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/230ff47c-ce58-4669-ba80-4a019365919d-logs\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.598074 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-config-data\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.598103 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d76v6\" (UniqueName: \"kubernetes.io/projected/2d0f44d4-a284-4c21-bea7-b964477081c4-kube-api-access-d76v6\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.598134 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.598554 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/230ff47c-ce58-4669-ba80-4a019365919d-logs\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.601527 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.603505 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.604216 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-config-data\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.609513 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.613954 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr2zd\" (UniqueName: \"kubernetes.io/projected/230ff47c-ce58-4669-ba80-4a019365919d-kube-api-access-fr2zd\") pod \"nova-api-0\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " pod="openstack/nova-api-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.614846 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d76v6\" (UniqueName: \"kubernetes.io/projected/2d0f44d4-a284-4c21-bea7-b964477081c4-kube-api-access-d76v6\") pod \"nova-cell1-conductor-0\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.667997 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.711611 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:39 crc kubenswrapper[4935]: I1005 08:52:39.757105 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.121692 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:40 crc kubenswrapper[4935]: W1005 08:52:40.133060 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c088042_dec3_4afd_918b_f0ca1615e88d.slice/crio-25f5c181260c924fb6094178f2fc7cda43919b3a07515c5103f0b96e95406b2c WatchSource:0}: Error finding container 25f5c181260c924fb6094178f2fc7cda43919b3a07515c5103f0b96e95406b2c: Status 404 returned error can't find the container with id 25f5c181260c924fb6094178f2fc7cda43919b3a07515c5103f0b96e95406b2c Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.194841 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c088042-dec3-4afd-918b-f0ca1615e88d","Type":"ContainerStarted","Data":"25f5c181260c924fb6094178f2fc7cda43919b3a07515c5103f0b96e95406b2c"} Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.198104 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"966eaf41-48ff-4110-bfab-eb465a2f2b0d","Type":"ContainerDied","Data":"86ebfe7621813106a1c14dbaa2e7f6ee9b0215e156a7f7750335b78564244abc"} Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.198147 4935 scope.go:117] "RemoveContainer" containerID="9284b372033818c061ffc34b2a5d36731698ac3d2fa520e0cafa6b3f5d158352" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.198251 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.290384 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.301427 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.316715 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.325680 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.347794 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.349343 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.352322 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.354648 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.414886 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-config-data\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.414965 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.415121 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnn66\" (UniqueName: \"kubernetes.io/projected/24bcc9c0-00ba-40dc-8f63-40068f01285d-kube-api-access-nnn66\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.475386 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.492537 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.517126 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnn66\" (UniqueName: \"kubernetes.io/projected/24bcc9c0-00ba-40dc-8f63-40068f01285d-kube-api-access-nnn66\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.517212 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-config-data\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.517249 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.521385 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.523054 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-config-data\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.538616 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnn66\" (UniqueName: \"kubernetes.io/projected/24bcc9c0-00ba-40dc-8f63-40068f01285d-kube-api-access-nnn66\") pod \"nova-scheduler-0\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.578109 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.653504 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d984f9d9c-s6xtd"] Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.653776 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" podUID="51a5af21-9df5-4364-924f-29386157ddcf" containerName="dnsmasq-dns" containerID="cri-o://1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317" gracePeriod=10 Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.668084 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.799586 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ce78cb3-073c-430a-8c06-a584144cfcd1" path="/var/lib/kubelet/pods/0ce78cb3-073c-430a-8c06-a584144cfcd1/volumes" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.805605 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9025cbb2-c30d-450b-bd98-ad68dd90d9e5" path="/var/lib/kubelet/pods/9025cbb2-c30d-450b-bd98-ad68dd90d9e5/volumes" Oct 05 08:52:40 crc kubenswrapper[4935]: I1005 08:52:40.806219 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="966eaf41-48ff-4110-bfab-eb465a2f2b0d" path="/var/lib/kubelet/pods/966eaf41-48ff-4110-bfab-eb465a2f2b0d/volumes" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.163501 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:41 crc kubenswrapper[4935]: W1005 08:52:41.186713 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24bcc9c0_00ba_40dc_8f63_40068f01285d.slice/crio-b2326c1d24ccc1fc3efbeb8cd31bce066152454db9d739c46e22a4ef875c7530 WatchSource:0}: Error finding container b2326c1d24ccc1fc3efbeb8cd31bce066152454db9d739c46e22a4ef875c7530: Status 404 returned error can't find the container with id b2326c1d24ccc1fc3efbeb8cd31bce066152454db9d739c46e22a4ef875c7530 Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.231772 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.232351 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"24bcc9c0-00ba-40dc-8f63-40068f01285d","Type":"ContainerStarted","Data":"b2326c1d24ccc1fc3efbeb8cd31bce066152454db9d739c46e22a4ef875c7530"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.239267 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2d0f44d4-a284-4c21-bea7-b964477081c4","Type":"ContainerStarted","Data":"d816192d2da7464a7d00b60d0018a5bea686b0f90a2a1722ee561a918945d8d7"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.239322 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2d0f44d4-a284-4c21-bea7-b964477081c4","Type":"ContainerStarted","Data":"d620ca9c46e083e75547ff906b742910a81e49c2215470687f193745cdf30b4a"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.239867 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.243506 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c088042-dec3-4afd-918b-f0ca1615e88d","Type":"ContainerStarted","Data":"438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.243652 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c088042-dec3-4afd-918b-f0ca1615e88d","Type":"ContainerStarted","Data":"54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.246072 4935 generic.go:334] "Generic (PLEG): container finished" podID="51a5af21-9df5-4364-924f-29386157ddcf" containerID="1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317" exitCode=0 Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.246204 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" event={"ID":"51a5af21-9df5-4364-924f-29386157ddcf","Type":"ContainerDied","Data":"1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.246297 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" event={"ID":"51a5af21-9df5-4364-924f-29386157ddcf","Type":"ContainerDied","Data":"ccd70b6826c0a988e2eac0d448274386c8c0de999f584db398719aac8b6ee280"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.246387 4935 scope.go:117] "RemoveContainer" containerID="1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.246590 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d984f9d9c-s6xtd" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.252445 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"230ff47c-ce58-4669-ba80-4a019365919d","Type":"ContainerStarted","Data":"827b26046055c9f9fd9d9e854fcb0daf5aed1ed65ad84a73bedae7416dd16b0b"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.252493 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"230ff47c-ce58-4669-ba80-4a019365919d","Type":"ContainerStarted","Data":"7728d67da2d6686025d477182f7bef9299baafa62206fb7d16b4fdcc5c76fd9c"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.252510 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"230ff47c-ce58-4669-ba80-4a019365919d","Type":"ContainerStarted","Data":"53c8bebb447698ceb9babfc38a959c96f084696b0f7f83a6fffd2e9f4e65cfa0"} Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.289051 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.289031106 podStartE2EDuration="2.289031106s" podCreationTimestamp="2025-10-05 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:41.283416237 +0000 UTC m=+7195.166042697" watchObservedRunningTime="2025-10-05 08:52:41.289031106 +0000 UTC m=+7195.171657566" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.296822 4935 scope.go:117] "RemoveContainer" containerID="220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.309276 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.328270 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.328250676 podStartE2EDuration="2.328250676s" podCreationTimestamp="2025-10-05 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:41.308650796 +0000 UTC m=+7195.191277256" watchObservedRunningTime="2025-10-05 08:52:41.328250676 +0000 UTC m=+7195.210877136" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.334299 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.334289816 podStartE2EDuration="2.334289816s" podCreationTimestamp="2025-10-05 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:41.326118399 +0000 UTC m=+7195.208744869" watchObservedRunningTime="2025-10-05 08:52:41.334289816 +0000 UTC m=+7195.216916276" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.336831 4935 scope.go:117] "RemoveContainer" containerID="1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317" Oct 05 08:52:41 crc kubenswrapper[4935]: E1005 08:52:41.337507 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317\": container with ID starting with 1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317 not found: ID does not exist" containerID="1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.337548 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317"} err="failed to get container status \"1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317\": rpc error: code = NotFound desc = could not find container \"1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317\": container with ID starting with 1a122b064c4969750d75fd2acb335e5779dec0d5434bc39390027f79eb435317 not found: ID does not exist" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.337574 4935 scope.go:117] "RemoveContainer" containerID="220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8" Oct 05 08:52:41 crc kubenswrapper[4935]: E1005 08:52:41.338042 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8\": container with ID starting with 220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8 not found: ID does not exist" containerID="220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.338077 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8"} err="failed to get container status \"220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8\": rpc error: code = NotFound desc = could not find container \"220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8\": container with ID starting with 220f4e68fb301354118a61bbd7967c352a9c27761346e6a42e8e4b745d6492e8 not found: ID does not exist" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.342198 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-config\") pod \"51a5af21-9df5-4364-924f-29386157ddcf\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.342286 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc\") pod \"51a5af21-9df5-4364-924f-29386157ddcf\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.342352 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-sb\") pod \"51a5af21-9df5-4364-924f-29386157ddcf\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.344817 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-nb\") pod \"51a5af21-9df5-4364-924f-29386157ddcf\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.344882 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpmbq\" (UniqueName: \"kubernetes.io/projected/51a5af21-9df5-4364-924f-29386157ddcf-kube-api-access-bpmbq\") pod \"51a5af21-9df5-4364-924f-29386157ddcf\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.377050 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a5af21-9df5-4364-924f-29386157ddcf-kube-api-access-bpmbq" (OuterVolumeSpecName: "kube-api-access-bpmbq") pod "51a5af21-9df5-4364-924f-29386157ddcf" (UID: "51a5af21-9df5-4364-924f-29386157ddcf"). InnerVolumeSpecName "kube-api-access-bpmbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.448285 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpmbq\" (UniqueName: \"kubernetes.io/projected/51a5af21-9df5-4364-924f-29386157ddcf-kube-api-access-bpmbq\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.449094 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-config" (OuterVolumeSpecName: "config") pod "51a5af21-9df5-4364-924f-29386157ddcf" (UID: "51a5af21-9df5-4364-924f-29386157ddcf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.452943 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "51a5af21-9df5-4364-924f-29386157ddcf" (UID: "51a5af21-9df5-4364-924f-29386157ddcf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:52:41 crc kubenswrapper[4935]: E1005 08:52:41.456132 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc podName:51a5af21-9df5-4364-924f-29386157ddcf nodeName:}" failed. No retries permitted until 2025-10-05 08:52:41.956101537 +0000 UTC m=+7195.838727997 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc") pod "51a5af21-9df5-4364-924f-29386157ddcf" (UID: "51a5af21-9df5-4364-924f-29386157ddcf") : error deleting /var/lib/kubelet/pods/51a5af21-9df5-4364-924f-29386157ddcf/volume-subpaths: remove /var/lib/kubelet/pods/51a5af21-9df5-4364-924f-29386157ddcf/volume-subpaths: no such file or directory Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.456349 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "51a5af21-9df5-4364-924f-29386157ddcf" (UID: "51a5af21-9df5-4364-924f-29386157ddcf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.549493 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.549888 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.549912 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.956450 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc\") pod \"51a5af21-9df5-4364-924f-29386157ddcf\" (UID: \"51a5af21-9df5-4364-924f-29386157ddcf\") " Oct 05 08:52:41 crc kubenswrapper[4935]: I1005 08:52:41.957483 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51a5af21-9df5-4364-924f-29386157ddcf" (UID: "51a5af21-9df5-4364-924f-29386157ddcf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:52:42 crc kubenswrapper[4935]: I1005 08:52:42.059529 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5af21-9df5-4364-924f-29386157ddcf-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:42 crc kubenswrapper[4935]: I1005 08:52:42.189356 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d984f9d9c-s6xtd"] Oct 05 08:52:42 crc kubenswrapper[4935]: I1005 08:52:42.213111 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d984f9d9c-s6xtd"] Oct 05 08:52:42 crc kubenswrapper[4935]: I1005 08:52:42.278268 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"24bcc9c0-00ba-40dc-8f63-40068f01285d","Type":"ContainerStarted","Data":"07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31"} Oct 05 08:52:42 crc kubenswrapper[4935]: I1005 08:52:42.304942 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.3049244030000002 podStartE2EDuration="2.304924403s" podCreationTimestamp="2025-10-05 08:52:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:42.298203015 +0000 UTC m=+7196.180829475" watchObservedRunningTime="2025-10-05 08:52:42.304924403 +0000 UTC m=+7196.187550863" Oct 05 08:52:42 crc kubenswrapper[4935]: I1005 08:52:42.790161 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a5af21-9df5-4364-924f-29386157ddcf" path="/var/lib/kubelet/pods/51a5af21-9df5-4364-924f-29386157ddcf/volumes" Oct 05 08:52:44 crc kubenswrapper[4935]: I1005 08:52:44.668255 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:52:44 crc kubenswrapper[4935]: I1005 08:52:44.668630 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:52:45 crc kubenswrapper[4935]: I1005 08:52:45.669121 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 08:52:49 crc kubenswrapper[4935]: I1005 08:52:49.669032 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:52:49 crc kubenswrapper[4935]: I1005 08:52:49.670946 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:52:49 crc kubenswrapper[4935]: I1005 08:52:49.756882 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 05 08:52:49 crc kubenswrapper[4935]: I1005 08:52:49.757138 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:52:49 crc kubenswrapper[4935]: I1005 08:52:49.757208 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:52:49 crc kubenswrapper[4935]: I1005 08:52:49.777339 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:52:49 crc kubenswrapper[4935]: E1005 08:52:49.777604 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.310395 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mvdz8"] Oct 05 08:52:50 crc kubenswrapper[4935]: E1005 08:52:50.311006 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a5af21-9df5-4364-924f-29386157ddcf" containerName="dnsmasq-dns" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.311023 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a5af21-9df5-4364-924f-29386157ddcf" containerName="dnsmasq-dns" Oct 05 08:52:50 crc kubenswrapper[4935]: E1005 08:52:50.311048 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a5af21-9df5-4364-924f-29386157ddcf" containerName="init" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.311054 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a5af21-9df5-4364-924f-29386157ddcf" containerName="init" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.311224 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a5af21-9df5-4364-924f-29386157ddcf" containerName="dnsmasq-dns" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.311817 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.319337 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.320104 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.323578 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mvdz8"] Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.413166 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-config-data\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.413313 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-scripts\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.413390 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6cz4\" (UniqueName: \"kubernetes.io/projected/1077b73f-70fa-4968-9ddb-4e2148bdea80-kube-api-access-g6cz4\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.413461 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.515006 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.515101 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-config-data\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.515215 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-scripts\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.515278 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6cz4\" (UniqueName: \"kubernetes.io/projected/1077b73f-70fa-4968-9ddb-4e2148bdea80-kube-api-access-g6cz4\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.525963 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-config-data\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.527240 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-scripts\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.530055 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.534613 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6cz4\" (UniqueName: \"kubernetes.io/projected/1077b73f-70fa-4968-9ddb-4e2148bdea80-kube-api-access-g6cz4\") pod \"nova-cell1-cell-mapping-mvdz8\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.637144 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.668865 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.726286 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.751115 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.78:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.752103 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.78:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.840156 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.80:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:52:50 crc kubenswrapper[4935]: I1005 08:52:50.840406 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.80:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:52:51 crc kubenswrapper[4935]: I1005 08:52:51.148461 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mvdz8"] Oct 05 08:52:51 crc kubenswrapper[4935]: I1005 08:52:51.372315 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mvdz8" event={"ID":"1077b73f-70fa-4968-9ddb-4e2148bdea80","Type":"ContainerStarted","Data":"13e3a91d48287ce623359e4ccfaba56866dc589067bb8fcf5ddb2de262b1282b"} Oct 05 08:52:51 crc kubenswrapper[4935]: I1005 08:52:51.372632 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mvdz8" event={"ID":"1077b73f-70fa-4968-9ddb-4e2148bdea80","Type":"ContainerStarted","Data":"1a1f76d9703549863813d4f5d1a22666f92dc6cf362c2eb70f057a439724d52e"} Oct 05 08:52:51 crc kubenswrapper[4935]: I1005 08:52:51.394624 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mvdz8" podStartSLOduration=1.3946088429999999 podStartE2EDuration="1.394608843s" podCreationTimestamp="2025-10-05 08:52:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:52:51.386820406 +0000 UTC m=+7205.269446866" watchObservedRunningTime="2025-10-05 08:52:51.394608843 +0000 UTC m=+7205.277235303" Oct 05 08:52:51 crc kubenswrapper[4935]: I1005 08:52:51.424656 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 08:52:56 crc kubenswrapper[4935]: I1005 08:52:56.430278 4935 generic.go:334] "Generic (PLEG): container finished" podID="1077b73f-70fa-4968-9ddb-4e2148bdea80" containerID="13e3a91d48287ce623359e4ccfaba56866dc589067bb8fcf5ddb2de262b1282b" exitCode=0 Oct 05 08:52:56 crc kubenswrapper[4935]: I1005 08:52:56.430319 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mvdz8" event={"ID":"1077b73f-70fa-4968-9ddb-4e2148bdea80","Type":"ContainerDied","Data":"13e3a91d48287ce623359e4ccfaba56866dc589067bb8fcf5ddb2de262b1282b"} Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.836604 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.883990 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-combined-ca-bundle\") pod \"1077b73f-70fa-4968-9ddb-4e2148bdea80\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.884083 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-scripts\") pod \"1077b73f-70fa-4968-9ddb-4e2148bdea80\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.884157 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-config-data\") pod \"1077b73f-70fa-4968-9ddb-4e2148bdea80\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.884378 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6cz4\" (UniqueName: \"kubernetes.io/projected/1077b73f-70fa-4968-9ddb-4e2148bdea80-kube-api-access-g6cz4\") pod \"1077b73f-70fa-4968-9ddb-4e2148bdea80\" (UID: \"1077b73f-70fa-4968-9ddb-4e2148bdea80\") " Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.889658 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1077b73f-70fa-4968-9ddb-4e2148bdea80-kube-api-access-g6cz4" (OuterVolumeSpecName: "kube-api-access-g6cz4") pod "1077b73f-70fa-4968-9ddb-4e2148bdea80" (UID: "1077b73f-70fa-4968-9ddb-4e2148bdea80"). InnerVolumeSpecName "kube-api-access-g6cz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.890469 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-scripts" (OuterVolumeSpecName: "scripts") pod "1077b73f-70fa-4968-9ddb-4e2148bdea80" (UID: "1077b73f-70fa-4968-9ddb-4e2148bdea80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.913025 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-config-data" (OuterVolumeSpecName: "config-data") pod "1077b73f-70fa-4968-9ddb-4e2148bdea80" (UID: "1077b73f-70fa-4968-9ddb-4e2148bdea80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.913862 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1077b73f-70fa-4968-9ddb-4e2148bdea80" (UID: "1077b73f-70fa-4968-9ddb-4e2148bdea80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.986053 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6cz4\" (UniqueName: \"kubernetes.io/projected/1077b73f-70fa-4968-9ddb-4e2148bdea80-kube-api-access-g6cz4\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.986088 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.986100 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:57 crc kubenswrapper[4935]: I1005 08:52:57.986111 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1077b73f-70fa-4968-9ddb-4e2148bdea80-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.454805 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mvdz8" event={"ID":"1077b73f-70fa-4968-9ddb-4e2148bdea80","Type":"ContainerDied","Data":"1a1f76d9703549863813d4f5d1a22666f92dc6cf362c2eb70f057a439724d52e"} Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.454865 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a1f76d9703549863813d4f5d1a22666f92dc6cf362c2eb70f057a439724d52e" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.454956 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mvdz8" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.667033 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v6zvv"] Oct 05 08:52:58 crc kubenswrapper[4935]: E1005 08:52:58.667962 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1077b73f-70fa-4968-9ddb-4e2148bdea80" containerName="nova-manage" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.668001 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1077b73f-70fa-4968-9ddb-4e2148bdea80" containerName="nova-manage" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.668410 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="1077b73f-70fa-4968-9ddb-4e2148bdea80" containerName="nova-manage" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.670779 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.679831 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.680283 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-log" containerID="cri-o://7728d67da2d6686025d477182f7bef9299baafa62206fb7d16b4fdcc5c76fd9c" gracePeriod=30 Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.680437 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-api" containerID="cri-o://827b26046055c9f9fd9d9e854fcb0daf5aed1ed65ad84a73bedae7416dd16b0b" gracePeriod=30 Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.689867 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6zvv"] Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.699239 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-catalog-content\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.699588 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-utilities\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.699780 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdcjc\" (UniqueName: \"kubernetes.io/projected/4de76672-bfe1-463b-be1c-547ed1a67beb-kube-api-access-xdcjc\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.804098 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-utilities\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.804222 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdcjc\" (UniqueName: \"kubernetes.io/projected/4de76672-bfe1-463b-be1c-547ed1a67beb-kube-api-access-xdcjc\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.804324 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-catalog-content\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.805369 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-utilities\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.807522 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-catalog-content\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.829963 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.830780 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="24bcc9c0-00ba-40dc-8f63-40068f01285d" containerName="nova-scheduler-scheduler" containerID="cri-o://07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31" gracePeriod=30 Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.841574 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.841990 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-log" containerID="cri-o://54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a" gracePeriod=30 Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.842166 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-metadata" containerID="cri-o://438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2" gracePeriod=30 Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.854574 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdcjc\" (UniqueName: \"kubernetes.io/projected/4de76672-bfe1-463b-be1c-547ed1a67beb-kube-api-access-xdcjc\") pod \"redhat-operators-v6zvv\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:58 crc kubenswrapper[4935]: I1005 08:52:58.997982 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:52:59 crc kubenswrapper[4935]: I1005 08:52:59.450467 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6zvv"] Oct 05 08:52:59 crc kubenswrapper[4935]: W1005 08:52:59.462183 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4de76672_bfe1_463b_be1c_547ed1a67beb.slice/crio-6caeb66587fbcc9d37cdcb8ca810f9e0328b348361f5dc2cfcb7472710076a81 WatchSource:0}: Error finding container 6caeb66587fbcc9d37cdcb8ca810f9e0328b348361f5dc2cfcb7472710076a81: Status 404 returned error can't find the container with id 6caeb66587fbcc9d37cdcb8ca810f9e0328b348361f5dc2cfcb7472710076a81 Oct 05 08:52:59 crc kubenswrapper[4935]: I1005 08:52:59.468445 4935 generic.go:334] "Generic (PLEG): container finished" podID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerID="54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a" exitCode=143 Oct 05 08:52:59 crc kubenswrapper[4935]: I1005 08:52:59.468512 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c088042-dec3-4afd-918b-f0ca1615e88d","Type":"ContainerDied","Data":"54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a"} Oct 05 08:52:59 crc kubenswrapper[4935]: I1005 08:52:59.474682 4935 generic.go:334] "Generic (PLEG): container finished" podID="230ff47c-ce58-4669-ba80-4a019365919d" containerID="7728d67da2d6686025d477182f7bef9299baafa62206fb7d16b4fdcc5c76fd9c" exitCode=143 Oct 05 08:52:59 crc kubenswrapper[4935]: I1005 08:52:59.474745 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"230ff47c-ce58-4669-ba80-4a019365919d","Type":"ContainerDied","Data":"7728d67da2d6686025d477182f7bef9299baafa62206fb7d16b4fdcc5c76fd9c"} Oct 05 08:53:00 crc kubenswrapper[4935]: I1005 08:53:00.486195 4935 generic.go:334] "Generic (PLEG): container finished" podID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerID="9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55" exitCode=0 Oct 05 08:53:00 crc kubenswrapper[4935]: I1005 08:53:00.486285 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zvv" event={"ID":"4de76672-bfe1-463b-be1c-547ed1a67beb","Type":"ContainerDied","Data":"9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55"} Oct 05 08:53:00 crc kubenswrapper[4935]: I1005 08:53:00.486718 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zvv" event={"ID":"4de76672-bfe1-463b-be1c-547ed1a67beb","Type":"ContainerStarted","Data":"6caeb66587fbcc9d37cdcb8ca810f9e0328b348361f5dc2cfcb7472710076a81"} Oct 05 08:53:00 crc kubenswrapper[4935]: I1005 08:53:00.488551 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:53:00 crc kubenswrapper[4935]: E1005 08:53:00.671824 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:53:00 crc kubenswrapper[4935]: E1005 08:53:00.673433 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:53:00 crc kubenswrapper[4935]: E1005 08:53:00.674806 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:53:00 crc kubenswrapper[4935]: E1005 08:53:00.674855 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="24bcc9c0-00ba-40dc-8f63-40068f01285d" containerName="nova-scheduler-scheduler" Oct 05 08:53:01 crc kubenswrapper[4935]: I1005 08:53:01.498074 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zvv" event={"ID":"4de76672-bfe1-463b-be1c-547ed1a67beb","Type":"ContainerStarted","Data":"1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9"} Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.504349 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.507328 4935 generic.go:334] "Generic (PLEG): container finished" podID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerID="438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2" exitCode=0 Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.507397 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.507412 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c088042-dec3-4afd-918b-f0ca1615e88d","Type":"ContainerDied","Data":"438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2"} Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.507464 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c088042-dec3-4afd-918b-f0ca1615e88d","Type":"ContainerDied","Data":"25f5c181260c924fb6094178f2fc7cda43919b3a07515c5103f0b96e95406b2c"} Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.507487 4935 scope.go:117] "RemoveContainer" containerID="438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.511133 4935 generic.go:334] "Generic (PLEG): container finished" podID="230ff47c-ce58-4669-ba80-4a019365919d" containerID="827b26046055c9f9fd9d9e854fcb0daf5aed1ed65ad84a73bedae7416dd16b0b" exitCode=0 Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.511189 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"230ff47c-ce58-4669-ba80-4a019365919d","Type":"ContainerDied","Data":"827b26046055c9f9fd9d9e854fcb0daf5aed1ed65ad84a73bedae7416dd16b0b"} Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.540698 4935 scope.go:117] "RemoveContainer" containerID="54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.559527 4935 scope.go:117] "RemoveContainer" containerID="438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2" Oct 05 08:53:02 crc kubenswrapper[4935]: E1005 08:53:02.560061 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2\": container with ID starting with 438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2 not found: ID does not exist" containerID="438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.560161 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2"} err="failed to get container status \"438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2\": rpc error: code = NotFound desc = could not find container \"438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2\": container with ID starting with 438f19d2f85ca5e58a15bd4e9279c55806f0ba0f7d9d10daabaf0d0cc69e83a2 not found: ID does not exist" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.560235 4935 scope.go:117] "RemoveContainer" containerID="54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a" Oct 05 08:53:02 crc kubenswrapper[4935]: E1005 08:53:02.560526 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a\": container with ID starting with 54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a not found: ID does not exist" containerID="54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.560555 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a"} err="failed to get container status \"54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a\": rpc error: code = NotFound desc = could not find container \"54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a\": container with ID starting with 54be4ebf7ae89e85be1f5768c9695a400fbc10522be75004bf00db24119d7c7a not found: ID does not exist" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.567108 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-combined-ca-bundle\") pod \"4c088042-dec3-4afd-918b-f0ca1615e88d\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.567232 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-config-data\") pod \"4c088042-dec3-4afd-918b-f0ca1615e88d\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.567324 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tphz2\" (UniqueName: \"kubernetes.io/projected/4c088042-dec3-4afd-918b-f0ca1615e88d-kube-api-access-tphz2\") pod \"4c088042-dec3-4afd-918b-f0ca1615e88d\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.567424 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c088042-dec3-4afd-918b-f0ca1615e88d-logs\") pod \"4c088042-dec3-4afd-918b-f0ca1615e88d\" (UID: \"4c088042-dec3-4afd-918b-f0ca1615e88d\") " Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.568040 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c088042-dec3-4afd-918b-f0ca1615e88d-logs" (OuterVolumeSpecName: "logs") pod "4c088042-dec3-4afd-918b-f0ca1615e88d" (UID: "4c088042-dec3-4afd-918b-f0ca1615e88d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.573950 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c088042-dec3-4afd-918b-f0ca1615e88d-kube-api-access-tphz2" (OuterVolumeSpecName: "kube-api-access-tphz2") pod "4c088042-dec3-4afd-918b-f0ca1615e88d" (UID: "4c088042-dec3-4afd-918b-f0ca1615e88d"). InnerVolumeSpecName "kube-api-access-tphz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.595022 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-config-data" (OuterVolumeSpecName: "config-data") pod "4c088042-dec3-4afd-918b-f0ca1615e88d" (UID: "4c088042-dec3-4afd-918b-f0ca1615e88d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.598950 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c088042-dec3-4afd-918b-f0ca1615e88d" (UID: "4c088042-dec3-4afd-918b-f0ca1615e88d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.668784 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.668817 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tphz2\" (UniqueName: \"kubernetes.io/projected/4c088042-dec3-4afd-918b-f0ca1615e88d-kube-api-access-tphz2\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.668826 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c088042-dec3-4afd-918b-f0ca1615e88d-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.668837 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c088042-dec3-4afd-918b-f0ca1615e88d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.843781 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.852503 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.869255 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:53:02 crc kubenswrapper[4935]: E1005 08:53:02.879201 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-metadata" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.879237 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-metadata" Oct 05 08:53:02 crc kubenswrapper[4935]: E1005 08:53:02.879510 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-log" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.879525 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-log" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.879736 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-metadata" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.881069 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" containerName="nova-metadata-log" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.882993 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.883126 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.888106 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.938295 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.977998 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-logs\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.978108 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fztsr\" (UniqueName: \"kubernetes.io/projected/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-kube-api-access-fztsr\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.978976 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-config-data\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:02 crc kubenswrapper[4935]: I1005 08:53:02.979071 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.080722 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr2zd\" (UniqueName: \"kubernetes.io/projected/230ff47c-ce58-4669-ba80-4a019365919d-kube-api-access-fr2zd\") pod \"230ff47c-ce58-4669-ba80-4a019365919d\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.080849 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/230ff47c-ce58-4669-ba80-4a019365919d-logs\") pod \"230ff47c-ce58-4669-ba80-4a019365919d\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.081372 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230ff47c-ce58-4669-ba80-4a019365919d-logs" (OuterVolumeSpecName: "logs") pod "230ff47c-ce58-4669-ba80-4a019365919d" (UID: "230ff47c-ce58-4669-ba80-4a019365919d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.081623 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-config-data\") pod \"230ff47c-ce58-4669-ba80-4a019365919d\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.082102 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-combined-ca-bundle\") pod \"230ff47c-ce58-4669-ba80-4a019365919d\" (UID: \"230ff47c-ce58-4669-ba80-4a019365919d\") " Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.082380 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fztsr\" (UniqueName: \"kubernetes.io/projected/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-kube-api-access-fztsr\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.082456 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-config-data\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.082498 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.082639 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-logs\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.082707 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/230ff47c-ce58-4669-ba80-4a019365919d-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.083254 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-logs\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.087188 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/230ff47c-ce58-4669-ba80-4a019365919d-kube-api-access-fr2zd" (OuterVolumeSpecName: "kube-api-access-fr2zd") pod "230ff47c-ce58-4669-ba80-4a019365919d" (UID: "230ff47c-ce58-4669-ba80-4a019365919d"). InnerVolumeSpecName "kube-api-access-fr2zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.089178 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-config-data\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.089399 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.101668 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fztsr\" (UniqueName: \"kubernetes.io/projected/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-kube-api-access-fztsr\") pod \"nova-metadata-0\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.126571 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-config-data" (OuterVolumeSpecName: "config-data") pod "230ff47c-ce58-4669-ba80-4a019365919d" (UID: "230ff47c-ce58-4669-ba80-4a019365919d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.128538 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "230ff47c-ce58-4669-ba80-4a019365919d" (UID: "230ff47c-ce58-4669-ba80-4a019365919d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.184340 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr2zd\" (UniqueName: \"kubernetes.io/projected/230ff47c-ce58-4669-ba80-4a019365919d-kube-api-access-fr2zd\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.184382 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.184396 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230ff47c-ce58-4669-ba80-4a019365919d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.235025 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.523172 4935 generic.go:334] "Generic (PLEG): container finished" podID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerID="1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9" exitCode=0 Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.523272 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zvv" event={"ID":"4de76672-bfe1-463b-be1c-547ed1a67beb","Type":"ContainerDied","Data":"1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9"} Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.526605 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.526644 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"230ff47c-ce58-4669-ba80-4a019365919d","Type":"ContainerDied","Data":"53c8bebb447698ceb9babfc38a959c96f084696b0f7f83a6fffd2e9f4e65cfa0"} Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.526700 4935 scope.go:117] "RemoveContainer" containerID="827b26046055c9f9fd9d9e854fcb0daf5aed1ed65ad84a73bedae7416dd16b0b" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.566919 4935 scope.go:117] "RemoveContainer" containerID="7728d67da2d6686025d477182f7bef9299baafa62206fb7d16b4fdcc5c76fd9c" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.575770 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.601404 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.609635 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 08:53:03 crc kubenswrapper[4935]: E1005 08:53:03.610121 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-api" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.610144 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-api" Oct 05 08:53:03 crc kubenswrapper[4935]: E1005 08:53:03.610171 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-log" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.610186 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-log" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.610454 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-log" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.610495 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="230ff47c-ce58-4669-ba80-4a019365919d" containerName="nova-api-api" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.611735 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.614001 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.621330 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.695314 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f473afb-6b80-4c81-83e3-140d713f354f-logs\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.695510 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-config-data\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.695550 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.695596 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwtx9\" (UniqueName: \"kubernetes.io/projected/2f473afb-6b80-4c81-83e3-140d713f354f-kube-api-access-vwtx9\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.756315 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.777305 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:53:03 crc kubenswrapper[4935]: E1005 08:53:03.777673 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.798945 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f473afb-6b80-4c81-83e3-140d713f354f-logs\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.799183 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-config-data\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.799273 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.799357 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwtx9\" (UniqueName: \"kubernetes.io/projected/2f473afb-6b80-4c81-83e3-140d713f354f-kube-api-access-vwtx9\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.799677 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f473afb-6b80-4c81-83e3-140d713f354f-logs\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.805068 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.805953 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-config-data\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.821484 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwtx9\" (UniqueName: \"kubernetes.io/projected/2f473afb-6b80-4c81-83e3-140d713f354f-kube-api-access-vwtx9\") pod \"nova-api-0\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " pod="openstack/nova-api-0" Oct 05 08:53:03 crc kubenswrapper[4935]: I1005 08:53:03.942633 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.050808 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.207591 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-config-data\") pod \"24bcc9c0-00ba-40dc-8f63-40068f01285d\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.208387 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-combined-ca-bundle\") pod \"24bcc9c0-00ba-40dc-8f63-40068f01285d\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.208433 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnn66\" (UniqueName: \"kubernetes.io/projected/24bcc9c0-00ba-40dc-8f63-40068f01285d-kube-api-access-nnn66\") pod \"24bcc9c0-00ba-40dc-8f63-40068f01285d\" (UID: \"24bcc9c0-00ba-40dc-8f63-40068f01285d\") " Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.215197 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24bcc9c0-00ba-40dc-8f63-40068f01285d-kube-api-access-nnn66" (OuterVolumeSpecName: "kube-api-access-nnn66") pod "24bcc9c0-00ba-40dc-8f63-40068f01285d" (UID: "24bcc9c0-00ba-40dc-8f63-40068f01285d"). InnerVolumeSpecName "kube-api-access-nnn66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.239262 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-config-data" (OuterVolumeSpecName: "config-data") pod "24bcc9c0-00ba-40dc-8f63-40068f01285d" (UID: "24bcc9c0-00ba-40dc-8f63-40068f01285d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.252856 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24bcc9c0-00ba-40dc-8f63-40068f01285d" (UID: "24bcc9c0-00ba-40dc-8f63-40068f01285d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.310908 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.310938 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnn66\" (UniqueName: \"kubernetes.io/projected/24bcc9c0-00ba-40dc-8f63-40068f01285d-kube-api-access-nnn66\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.310951 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24bcc9c0-00ba-40dc-8f63-40068f01285d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.475240 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.567260 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f473afb-6b80-4c81-83e3-140d713f354f","Type":"ContainerStarted","Data":"4a7a28e72b7aee223e0d31f407b2416e2db85e3b7564f74358ce3b66c565ff60"} Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.570156 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1","Type":"ContainerStarted","Data":"28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e"} Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.570199 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1","Type":"ContainerStarted","Data":"778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3"} Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.570215 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1","Type":"ContainerStarted","Data":"9e4e0e430cd7be262998c776883ae0e2a5807572e21e394c6812e78703b2f431"} Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.587737 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zvv" event={"ID":"4de76672-bfe1-463b-be1c-547ed1a67beb","Type":"ContainerStarted","Data":"6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a"} Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.595618 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.595602398 podStartE2EDuration="2.595602398s" podCreationTimestamp="2025-10-05 08:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:53:04.595102545 +0000 UTC m=+7218.477729005" watchObservedRunningTime="2025-10-05 08:53:04.595602398 +0000 UTC m=+7218.478228868" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.596101 4935 generic.go:334] "Generic (PLEG): container finished" podID="24bcc9c0-00ba-40dc-8f63-40068f01285d" containerID="07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31" exitCode=0 Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.596151 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"24bcc9c0-00ba-40dc-8f63-40068f01285d","Type":"ContainerDied","Data":"07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31"} Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.596178 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"24bcc9c0-00ba-40dc-8f63-40068f01285d","Type":"ContainerDied","Data":"b2326c1d24ccc1fc3efbeb8cd31bce066152454db9d739c46e22a4ef875c7530"} Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.596195 4935 scope.go:117] "RemoveContainer" containerID="07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.596325 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.629352 4935 scope.go:117] "RemoveContainer" containerID="07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31" Oct 05 08:53:04 crc kubenswrapper[4935]: E1005 08:53:04.632442 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31\": container with ID starting with 07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31 not found: ID does not exist" containerID="07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.632498 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31"} err="failed to get container status \"07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31\": rpc error: code = NotFound desc = could not find container \"07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31\": container with ID starting with 07a5904290ecbdee1d536260340d056a9aba62dd7a51c159698b25bfda612c31 not found: ID does not exist" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.648030 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v6zvv" podStartSLOduration=3.151572243 podStartE2EDuration="6.647996368s" podCreationTimestamp="2025-10-05 08:52:58 +0000 UTC" firstStartedPulling="2025-10-05 08:53:00.48834706 +0000 UTC m=+7214.370973520" lastFinishedPulling="2025-10-05 08:53:03.984771185 +0000 UTC m=+7217.867397645" observedRunningTime="2025-10-05 08:53:04.618060654 +0000 UTC m=+7218.500687134" watchObservedRunningTime="2025-10-05 08:53:04.647996368 +0000 UTC m=+7218.530622828" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.657595 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.672706 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.690598 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:53:04 crc kubenswrapper[4935]: E1005 08:53:04.691102 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24bcc9c0-00ba-40dc-8f63-40068f01285d" containerName="nova-scheduler-scheduler" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.691124 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="24bcc9c0-00ba-40dc-8f63-40068f01285d" containerName="nova-scheduler-scheduler" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.691334 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="24bcc9c0-00ba-40dc-8f63-40068f01285d" containerName="nova-scheduler-scheduler" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.692085 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.694381 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.697831 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.790202 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="230ff47c-ce58-4669-ba80-4a019365919d" path="/var/lib/kubelet/pods/230ff47c-ce58-4669-ba80-4a019365919d/volumes" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.791370 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24bcc9c0-00ba-40dc-8f63-40068f01285d" path="/var/lib/kubelet/pods/24bcc9c0-00ba-40dc-8f63-40068f01285d/volumes" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.793338 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c088042-dec3-4afd-918b-f0ca1615e88d" path="/var/lib/kubelet/pods/4c088042-dec3-4afd-918b-f0ca1615e88d/volumes" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.831955 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87x8\" (UniqueName: \"kubernetes.io/projected/49137eff-8dd5-4896-9770-77a953079e83-kube-api-access-q87x8\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.832072 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.832123 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-config-data\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.934271 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-config-data\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.934447 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87x8\" (UniqueName: \"kubernetes.io/projected/49137eff-8dd5-4896-9770-77a953079e83-kube-api-access-q87x8\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.934568 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.939593 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.939703 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-config-data\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:04 crc kubenswrapper[4935]: I1005 08:53:04.951508 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87x8\" (UniqueName: \"kubernetes.io/projected/49137eff-8dd5-4896-9770-77a953079e83-kube-api-access-q87x8\") pod \"nova-scheduler-0\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " pod="openstack/nova-scheduler-0" Oct 05 08:53:05 crc kubenswrapper[4935]: I1005 08:53:05.006859 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:53:05 crc kubenswrapper[4935]: I1005 08:53:05.419149 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:53:05 crc kubenswrapper[4935]: W1005 08:53:05.428451 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49137eff_8dd5_4896_9770_77a953079e83.slice/crio-280504ca9e05db6128fa92fb6187e0e0ebe929882ce918221a6b5cdb501058ee WatchSource:0}: Error finding container 280504ca9e05db6128fa92fb6187e0e0ebe929882ce918221a6b5cdb501058ee: Status 404 returned error can't find the container with id 280504ca9e05db6128fa92fb6187e0e0ebe929882ce918221a6b5cdb501058ee Oct 05 08:53:05 crc kubenswrapper[4935]: I1005 08:53:05.614381 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"49137eff-8dd5-4896-9770-77a953079e83","Type":"ContainerStarted","Data":"280504ca9e05db6128fa92fb6187e0e0ebe929882ce918221a6b5cdb501058ee"} Oct 05 08:53:05 crc kubenswrapper[4935]: I1005 08:53:05.618921 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f473afb-6b80-4c81-83e3-140d713f354f","Type":"ContainerStarted","Data":"5784c136d0f06f0ed1fa0e7636182c1d63a5241795a6a516c6f12baf39969bd5"} Oct 05 08:53:05 crc kubenswrapper[4935]: I1005 08:53:05.620174 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f473afb-6b80-4c81-83e3-140d713f354f","Type":"ContainerStarted","Data":"fe12d4a3ada614e73e199d1a7d8f93d5fb49ef898df4b349402244188a337eb9"} Oct 05 08:53:05 crc kubenswrapper[4935]: I1005 08:53:05.645243 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.64521888 podStartE2EDuration="2.64521888s" podCreationTimestamp="2025-10-05 08:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:53:05.635619756 +0000 UTC m=+7219.518246226" watchObservedRunningTime="2025-10-05 08:53:05.64521888 +0000 UTC m=+7219.527845340" Oct 05 08:53:06 crc kubenswrapper[4935]: I1005 08:53:06.629528 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"49137eff-8dd5-4896-9770-77a953079e83","Type":"ContainerStarted","Data":"82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87"} Oct 05 08:53:06 crc kubenswrapper[4935]: I1005 08:53:06.647119 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.647101515 podStartE2EDuration="2.647101515s" podCreationTimestamp="2025-10-05 08:53:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:53:06.642815151 +0000 UTC m=+7220.525441611" watchObservedRunningTime="2025-10-05 08:53:06.647101515 +0000 UTC m=+7220.529727975" Oct 05 08:53:08 crc kubenswrapper[4935]: I1005 08:53:08.235860 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:53:08 crc kubenswrapper[4935]: I1005 08:53:08.236201 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:53:08 crc kubenswrapper[4935]: I1005 08:53:08.998796 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:53:08 crc kubenswrapper[4935]: I1005 08:53:08.998866 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:53:09 crc kubenswrapper[4935]: I1005 08:53:09.074186 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:53:09 crc kubenswrapper[4935]: I1005 08:53:09.714430 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:53:10 crc kubenswrapper[4935]: I1005 08:53:10.009044 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 08:53:11 crc kubenswrapper[4935]: I1005 08:53:11.638271 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v6zvv"] Oct 05 08:53:11 crc kubenswrapper[4935]: I1005 08:53:11.684139 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v6zvv" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerName="registry-server" containerID="cri-o://6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a" gracePeriod=2 Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.161623 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.189887 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-catalog-content\") pod \"4de76672-bfe1-463b-be1c-547ed1a67beb\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.190177 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-utilities\") pod \"4de76672-bfe1-463b-be1c-547ed1a67beb\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.190267 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdcjc\" (UniqueName: \"kubernetes.io/projected/4de76672-bfe1-463b-be1c-547ed1a67beb-kube-api-access-xdcjc\") pod \"4de76672-bfe1-463b-be1c-547ed1a67beb\" (UID: \"4de76672-bfe1-463b-be1c-547ed1a67beb\") " Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.193473 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-utilities" (OuterVolumeSpecName: "utilities") pod "4de76672-bfe1-463b-be1c-547ed1a67beb" (UID: "4de76672-bfe1-463b-be1c-547ed1a67beb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.200644 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4de76672-bfe1-463b-be1c-547ed1a67beb-kube-api-access-xdcjc" (OuterVolumeSpecName: "kube-api-access-xdcjc") pod "4de76672-bfe1-463b-be1c-547ed1a67beb" (UID: "4de76672-bfe1-463b-be1c-547ed1a67beb"). InnerVolumeSpecName "kube-api-access-xdcjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.279039 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4de76672-bfe1-463b-be1c-547ed1a67beb" (UID: "4de76672-bfe1-463b-be1c-547ed1a67beb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.291692 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.291726 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdcjc\" (UniqueName: \"kubernetes.io/projected/4de76672-bfe1-463b-be1c-547ed1a67beb-kube-api-access-xdcjc\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.291735 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de76672-bfe1-463b-be1c-547ed1a67beb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.703034 4935 generic.go:334] "Generic (PLEG): container finished" podID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerID="6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a" exitCode=0 Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.703094 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zvv" event={"ID":"4de76672-bfe1-463b-be1c-547ed1a67beb","Type":"ContainerDied","Data":"6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a"} Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.703156 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6zvv" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.703189 4935 scope.go:117] "RemoveContainer" containerID="6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.703168 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zvv" event={"ID":"4de76672-bfe1-463b-be1c-547ed1a67beb","Type":"ContainerDied","Data":"6caeb66587fbcc9d37cdcb8ca810f9e0328b348361f5dc2cfcb7472710076a81"} Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.733941 4935 scope.go:117] "RemoveContainer" containerID="1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.752398 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v6zvv"] Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.768472 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v6zvv"] Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.776603 4935 scope.go:117] "RemoveContainer" containerID="9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.793324 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" path="/var/lib/kubelet/pods/4de76672-bfe1-463b-be1c-547ed1a67beb/volumes" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.833076 4935 scope.go:117] "RemoveContainer" containerID="6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a" Oct 05 08:53:12 crc kubenswrapper[4935]: E1005 08:53:12.833723 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a\": container with ID starting with 6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a not found: ID does not exist" containerID="6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.833786 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a"} err="failed to get container status \"6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a\": rpc error: code = NotFound desc = could not find container \"6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a\": container with ID starting with 6052a56f9d9f81e31f730c67942d18ae372e82c51353419123f8a97b2fe93f3a not found: ID does not exist" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.833820 4935 scope.go:117] "RemoveContainer" containerID="1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9" Oct 05 08:53:12 crc kubenswrapper[4935]: E1005 08:53:12.834204 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9\": container with ID starting with 1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9 not found: ID does not exist" containerID="1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.834248 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9"} err="failed to get container status \"1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9\": rpc error: code = NotFound desc = could not find container \"1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9\": container with ID starting with 1a0507f41108ba4006daf98af4c712d6cfa0a01ebba5135f2398ef01fffc46b9 not found: ID does not exist" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.834265 4935 scope.go:117] "RemoveContainer" containerID="9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55" Oct 05 08:53:12 crc kubenswrapper[4935]: E1005 08:53:12.834639 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55\": container with ID starting with 9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55 not found: ID does not exist" containerID="9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55" Oct 05 08:53:12 crc kubenswrapper[4935]: I1005 08:53:12.834662 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55"} err="failed to get container status \"9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55\": rpc error: code = NotFound desc = could not find container \"9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55\": container with ID starting with 9f818c08255efb049d63040d954c4cb0ed229291e497375052b31b11e1b8ec55 not found: ID does not exist" Oct 05 08:53:13 crc kubenswrapper[4935]: I1005 08:53:13.235718 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:53:13 crc kubenswrapper[4935]: I1005 08:53:13.236037 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:53:13 crc kubenswrapper[4935]: I1005 08:53:13.944017 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:53:13 crc kubenswrapper[4935]: I1005 08:53:13.944523 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:53:14 crc kubenswrapper[4935]: I1005 08:53:14.330057 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:53:14 crc kubenswrapper[4935]: I1005 08:53:14.330439 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:53:15 crc kubenswrapper[4935]: I1005 08:53:15.007266 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 08:53:15 crc kubenswrapper[4935]: I1005 08:53:15.027260 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.85:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:53:15 crc kubenswrapper[4935]: I1005 08:53:15.027582 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.85:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:53:15 crc kubenswrapper[4935]: I1005 08:53:15.053051 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 08:53:15 crc kubenswrapper[4935]: I1005 08:53:15.773124 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 08:53:18 crc kubenswrapper[4935]: I1005 08:53:18.777847 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:53:18 crc kubenswrapper[4935]: E1005 08:53:18.778783 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:53:23 crc kubenswrapper[4935]: I1005 08:53:23.239575 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 08:53:23 crc kubenswrapper[4935]: I1005 08:53:23.244820 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 08:53:23 crc kubenswrapper[4935]: I1005 08:53:23.245587 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 08:53:23 crc kubenswrapper[4935]: I1005 08:53:23.808580 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 08:53:23 crc kubenswrapper[4935]: I1005 08:53:23.956290 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 08:53:23 crc kubenswrapper[4935]: I1005 08:53:23.957258 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 08:53:23 crc kubenswrapper[4935]: I1005 08:53:23.958342 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:53:23 crc kubenswrapper[4935]: I1005 08:53:23.967489 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 08:53:24 crc kubenswrapper[4935]: I1005 08:53:24.815469 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:53:24 crc kubenswrapper[4935]: I1005 08:53:24.820563 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.044533 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-648568c79-jw292"] Oct 05 08:53:25 crc kubenswrapper[4935]: E1005 08:53:25.044901 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerName="extract-utilities" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.044920 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerName="extract-utilities" Oct 05 08:53:25 crc kubenswrapper[4935]: E1005 08:53:25.044934 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerName="extract-content" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.044940 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerName="extract-content" Oct 05 08:53:25 crc kubenswrapper[4935]: E1005 08:53:25.044968 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerName="registry-server" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.044975 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerName="registry-server" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.045144 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="4de76672-bfe1-463b-be1c-547ed1a67beb" containerName="registry-server" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.046204 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.059755 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-648568c79-jw292"] Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.146601 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-config\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.146984 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7slr8\" (UniqueName: \"kubernetes.io/projected/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-kube-api-access-7slr8\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.147013 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-sb\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.147068 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-dns-svc\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.147089 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-nb\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.248847 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7slr8\" (UniqueName: \"kubernetes.io/projected/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-kube-api-access-7slr8\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.249144 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-sb\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.249273 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-dns-svc\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.249384 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-nb\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.249541 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-config\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.250348 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-config\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.250374 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-sb\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.251045 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-nb\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.251054 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-dns-svc\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.287316 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7slr8\" (UniqueName: \"kubernetes.io/projected/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-kube-api-access-7slr8\") pod \"dnsmasq-dns-648568c79-jw292\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.387039 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:25 crc kubenswrapper[4935]: I1005 08:53:25.869602 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-648568c79-jw292"] Oct 05 08:53:26 crc kubenswrapper[4935]: I1005 08:53:26.831596 4935 generic.go:334] "Generic (PLEG): container finished" podID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" containerID="7fd479a665b838529e89e194e7b197f88ec85fa155ef4597f8094eddd230e14f" exitCode=0 Oct 05 08:53:26 crc kubenswrapper[4935]: I1005 08:53:26.831880 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648568c79-jw292" event={"ID":"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d","Type":"ContainerDied","Data":"7fd479a665b838529e89e194e7b197f88ec85fa155ef4597f8094eddd230e14f"} Oct 05 08:53:26 crc kubenswrapper[4935]: I1005 08:53:26.832129 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648568c79-jw292" event={"ID":"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d","Type":"ContainerStarted","Data":"310f2c1409d80af8c81ad8f8a78f6533a7831b8cb2eb140683e829acf9857407"} Oct 05 08:53:27 crc kubenswrapper[4935]: I1005 08:53:27.841791 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648568c79-jw292" event={"ID":"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d","Type":"ContainerStarted","Data":"058b4fb7475b58ff5899f48eaa3828f618db23c42f22f67cd913e50c3b6f87d2"} Oct 05 08:53:27 crc kubenswrapper[4935]: I1005 08:53:27.843251 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:27 crc kubenswrapper[4935]: I1005 08:53:27.863283 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-648568c79-jw292" podStartSLOduration=2.863268098 podStartE2EDuration="2.863268098s" podCreationTimestamp="2025-10-05 08:53:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:53:27.861718437 +0000 UTC m=+7241.744344897" watchObservedRunningTime="2025-10-05 08:53:27.863268098 +0000 UTC m=+7241.745894558" Oct 05 08:53:29 crc kubenswrapper[4935]: I1005 08:53:29.778796 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:53:29 crc kubenswrapper[4935]: E1005 08:53:29.779484 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:53:35 crc kubenswrapper[4935]: I1005 08:53:35.389687 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:53:35 crc kubenswrapper[4935]: I1005 08:53:35.458032 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84dc58f955-ljs2d"] Oct 05 08:53:35 crc kubenswrapper[4935]: I1005 08:53:35.458371 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" podUID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerName="dnsmasq-dns" containerID="cri-o://1dc48e927d06f30f333ea6f79b420948c1d413f840ac9cfa35e9cb1568cc09f7" gracePeriod=10 Oct 05 08:53:35 crc kubenswrapper[4935]: I1005 08:53:35.576299 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" podUID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.75:5353: connect: connection refused" Oct 05 08:53:35 crc kubenswrapper[4935]: I1005 08:53:35.924809 4935 generic.go:334] "Generic (PLEG): container finished" podID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerID="1dc48e927d06f30f333ea6f79b420948c1d413f840ac9cfa35e9cb1568cc09f7" exitCode=0 Oct 05 08:53:35 crc kubenswrapper[4935]: I1005 08:53:35.925127 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" event={"ID":"eb2cfe55-91a7-4893-839f-64b3d6d46c9f","Type":"ContainerDied","Data":"1dc48e927d06f30f333ea6f79b420948c1d413f840ac9cfa35e9cb1568cc09f7"} Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.020100 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.035209 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-nb\") pod \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.035297 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-sb\") pod \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.035347 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf8rx\" (UniqueName: \"kubernetes.io/projected/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-kube-api-access-pf8rx\") pod \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.035400 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-dns-svc\") pod \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.035492 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-config\") pod \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\" (UID: \"eb2cfe55-91a7-4893-839f-64b3d6d46c9f\") " Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.046582 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-kube-api-access-pf8rx" (OuterVolumeSpecName: "kube-api-access-pf8rx") pod "eb2cfe55-91a7-4893-839f-64b3d6d46c9f" (UID: "eb2cfe55-91a7-4893-839f-64b3d6d46c9f"). InnerVolumeSpecName "kube-api-access-pf8rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.095012 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-config" (OuterVolumeSpecName: "config") pod "eb2cfe55-91a7-4893-839f-64b3d6d46c9f" (UID: "eb2cfe55-91a7-4893-839f-64b3d6d46c9f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.107929 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb2cfe55-91a7-4893-839f-64b3d6d46c9f" (UID: "eb2cfe55-91a7-4893-839f-64b3d6d46c9f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.108212 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb2cfe55-91a7-4893-839f-64b3d6d46c9f" (UID: "eb2cfe55-91a7-4893-839f-64b3d6d46c9f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.114313 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb2cfe55-91a7-4893-839f-64b3d6d46c9f" (UID: "eb2cfe55-91a7-4893-839f-64b3d6d46c9f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.142180 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.142235 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.142246 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.142256 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf8rx\" (UniqueName: \"kubernetes.io/projected/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-kube-api-access-pf8rx\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.142264 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb2cfe55-91a7-4893-839f-64b3d6d46c9f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.939187 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" event={"ID":"eb2cfe55-91a7-4893-839f-64b3d6d46c9f","Type":"ContainerDied","Data":"346548fa1c79c59699de1fb827fd30f47a0d739ea16ab3521bffb799513e5c67"} Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.939289 4935 scope.go:117] "RemoveContainer" containerID="1dc48e927d06f30f333ea6f79b420948c1d413f840ac9cfa35e9cb1568cc09f7" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.939519 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84dc58f955-ljs2d" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.982598 4935 scope.go:117] "RemoveContainer" containerID="8523e9c5ce4cde1a4397f2e533b4c940d0d1709fad8441355e1d552c5fc1aa7f" Oct 05 08:53:36 crc kubenswrapper[4935]: I1005 08:53:36.986411 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84dc58f955-ljs2d"] Oct 05 08:53:37 crc kubenswrapper[4935]: I1005 08:53:37.012752 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84dc58f955-ljs2d"] Oct 05 08:53:38 crc kubenswrapper[4935]: I1005 08:53:38.791286 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" path="/var/lib/kubelet/pods/eb2cfe55-91a7-4893-839f-64b3d6d46c9f/volumes" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.451267 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-lp2bz"] Oct 05 08:53:39 crc kubenswrapper[4935]: E1005 08:53:39.453999 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerName="init" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.454027 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerName="init" Oct 05 08:53:39 crc kubenswrapper[4935]: E1005 08:53:39.454052 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerName="dnsmasq-dns" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.454060 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerName="dnsmasq-dns" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.454253 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb2cfe55-91a7-4893-839f-64b3d6d46c9f" containerName="dnsmasq-dns" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.454824 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lp2bz" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.460196 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lp2bz"] Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.504397 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jlvs\" (UniqueName: \"kubernetes.io/projected/a680f5ef-08d9-4f31-a4a5-5230086909a6-kube-api-access-6jlvs\") pod \"cinder-db-create-lp2bz\" (UID: \"a680f5ef-08d9-4f31-a4a5-5230086909a6\") " pod="openstack/cinder-db-create-lp2bz" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.605459 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jlvs\" (UniqueName: \"kubernetes.io/projected/a680f5ef-08d9-4f31-a4a5-5230086909a6-kube-api-access-6jlvs\") pod \"cinder-db-create-lp2bz\" (UID: \"a680f5ef-08d9-4f31-a4a5-5230086909a6\") " pod="openstack/cinder-db-create-lp2bz" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.637666 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jlvs\" (UniqueName: \"kubernetes.io/projected/a680f5ef-08d9-4f31-a4a5-5230086909a6-kube-api-access-6jlvs\") pod \"cinder-db-create-lp2bz\" (UID: \"a680f5ef-08d9-4f31-a4a5-5230086909a6\") " pod="openstack/cinder-db-create-lp2bz" Oct 05 08:53:39 crc kubenswrapper[4935]: I1005 08:53:39.818848 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lp2bz" Oct 05 08:53:40 crc kubenswrapper[4935]: W1005 08:53:40.270185 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda680f5ef_08d9_4f31_a4a5_5230086909a6.slice/crio-c75126fb7a3083627730a94b979d7cb16ca70526015378dd93d98d92901042a9 WatchSource:0}: Error finding container c75126fb7a3083627730a94b979d7cb16ca70526015378dd93d98d92901042a9: Status 404 returned error can't find the container with id c75126fb7a3083627730a94b979d7cb16ca70526015378dd93d98d92901042a9 Oct 05 08:53:40 crc kubenswrapper[4935]: I1005 08:53:40.280119 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lp2bz"] Oct 05 08:53:40 crc kubenswrapper[4935]: I1005 08:53:40.979276 4935 generic.go:334] "Generic (PLEG): container finished" podID="a680f5ef-08d9-4f31-a4a5-5230086909a6" containerID="360fa3306cc49a1bdddcc86ff46c7160a777c9ee86c27f808dc4b62093559767" exitCode=0 Oct 05 08:53:40 crc kubenswrapper[4935]: I1005 08:53:40.979427 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lp2bz" event={"ID":"a680f5ef-08d9-4f31-a4a5-5230086909a6","Type":"ContainerDied","Data":"360fa3306cc49a1bdddcc86ff46c7160a777c9ee86c27f808dc4b62093559767"} Oct 05 08:53:40 crc kubenswrapper[4935]: I1005 08:53:40.979802 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lp2bz" event={"ID":"a680f5ef-08d9-4f31-a4a5-5230086909a6","Type":"ContainerStarted","Data":"c75126fb7a3083627730a94b979d7cb16ca70526015378dd93d98d92901042a9"} Oct 05 08:53:42 crc kubenswrapper[4935]: I1005 08:53:42.421513 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lp2bz" Oct 05 08:53:42 crc kubenswrapper[4935]: I1005 08:53:42.464805 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jlvs\" (UniqueName: \"kubernetes.io/projected/a680f5ef-08d9-4f31-a4a5-5230086909a6-kube-api-access-6jlvs\") pod \"a680f5ef-08d9-4f31-a4a5-5230086909a6\" (UID: \"a680f5ef-08d9-4f31-a4a5-5230086909a6\") " Oct 05 08:53:42 crc kubenswrapper[4935]: I1005 08:53:42.472638 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a680f5ef-08d9-4f31-a4a5-5230086909a6-kube-api-access-6jlvs" (OuterVolumeSpecName: "kube-api-access-6jlvs") pod "a680f5ef-08d9-4f31-a4a5-5230086909a6" (UID: "a680f5ef-08d9-4f31-a4a5-5230086909a6"). InnerVolumeSpecName "kube-api-access-6jlvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:53:42 crc kubenswrapper[4935]: I1005 08:53:42.567873 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jlvs\" (UniqueName: \"kubernetes.io/projected/a680f5ef-08d9-4f31-a4a5-5230086909a6-kube-api-access-6jlvs\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:42 crc kubenswrapper[4935]: I1005 08:53:42.779432 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:53:42 crc kubenswrapper[4935]: E1005 08:53:42.779934 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:53:43 crc kubenswrapper[4935]: I1005 08:53:43.007509 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lp2bz" event={"ID":"a680f5ef-08d9-4f31-a4a5-5230086909a6","Type":"ContainerDied","Data":"c75126fb7a3083627730a94b979d7cb16ca70526015378dd93d98d92901042a9"} Oct 05 08:53:43 crc kubenswrapper[4935]: I1005 08:53:43.007566 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c75126fb7a3083627730a94b979d7cb16ca70526015378dd93d98d92901042a9" Oct 05 08:53:43 crc kubenswrapper[4935]: I1005 08:53:43.007587 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lp2bz" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.580953 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-fd41-account-create-nkmgw"] Oct 05 08:53:49 crc kubenswrapper[4935]: E1005 08:53:49.582717 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a680f5ef-08d9-4f31-a4a5-5230086909a6" containerName="mariadb-database-create" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.582751 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a680f5ef-08d9-4f31-a4a5-5230086909a6" containerName="mariadb-database-create" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.583428 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a680f5ef-08d9-4f31-a4a5-5230086909a6" containerName="mariadb-database-create" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.584825 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fd41-account-create-nkmgw" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.589119 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.603477 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-fd41-account-create-nkmgw"] Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.712179 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnqj5\" (UniqueName: \"kubernetes.io/projected/313db7f6-e689-4071-87d8-71ea5bc7cb30-kube-api-access-fnqj5\") pod \"cinder-fd41-account-create-nkmgw\" (UID: \"313db7f6-e689-4071-87d8-71ea5bc7cb30\") " pod="openstack/cinder-fd41-account-create-nkmgw" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.814607 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnqj5\" (UniqueName: \"kubernetes.io/projected/313db7f6-e689-4071-87d8-71ea5bc7cb30-kube-api-access-fnqj5\") pod \"cinder-fd41-account-create-nkmgw\" (UID: \"313db7f6-e689-4071-87d8-71ea5bc7cb30\") " pod="openstack/cinder-fd41-account-create-nkmgw" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.856520 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnqj5\" (UniqueName: \"kubernetes.io/projected/313db7f6-e689-4071-87d8-71ea5bc7cb30-kube-api-access-fnqj5\") pod \"cinder-fd41-account-create-nkmgw\" (UID: \"313db7f6-e689-4071-87d8-71ea5bc7cb30\") " pod="openstack/cinder-fd41-account-create-nkmgw" Oct 05 08:53:49 crc kubenswrapper[4935]: I1005 08:53:49.950984 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fd41-account-create-nkmgw" Oct 05 08:53:50 crc kubenswrapper[4935]: W1005 08:53:50.431034 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod313db7f6_e689_4071_87d8_71ea5bc7cb30.slice/crio-94aef2e9c7d9e7eb14f95f9ea32feb6c6b176acb20c8b606afddb25f6c52c96c WatchSource:0}: Error finding container 94aef2e9c7d9e7eb14f95f9ea32feb6c6b176acb20c8b606afddb25f6c52c96c: Status 404 returned error can't find the container with id 94aef2e9c7d9e7eb14f95f9ea32feb6c6b176acb20c8b606afddb25f6c52c96c Oct 05 08:53:50 crc kubenswrapper[4935]: I1005 08:53:50.434374 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-fd41-account-create-nkmgw"] Oct 05 08:53:51 crc kubenswrapper[4935]: I1005 08:53:51.084519 4935 generic.go:334] "Generic (PLEG): container finished" podID="313db7f6-e689-4071-87d8-71ea5bc7cb30" containerID="128166db61018ffddcb331e01c51278b845e221547a789749ca241d6a183dd99" exitCode=0 Oct 05 08:53:51 crc kubenswrapper[4935]: I1005 08:53:51.084613 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fd41-account-create-nkmgw" event={"ID":"313db7f6-e689-4071-87d8-71ea5bc7cb30","Type":"ContainerDied","Data":"128166db61018ffddcb331e01c51278b845e221547a789749ca241d6a183dd99"} Oct 05 08:53:51 crc kubenswrapper[4935]: I1005 08:53:51.084804 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fd41-account-create-nkmgw" event={"ID":"313db7f6-e689-4071-87d8-71ea5bc7cb30","Type":"ContainerStarted","Data":"94aef2e9c7d9e7eb14f95f9ea32feb6c6b176acb20c8b606afddb25f6c52c96c"} Oct 05 08:53:52 crc kubenswrapper[4935]: I1005 08:53:52.557366 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fd41-account-create-nkmgw" Oct 05 08:53:52 crc kubenswrapper[4935]: I1005 08:53:52.667477 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnqj5\" (UniqueName: \"kubernetes.io/projected/313db7f6-e689-4071-87d8-71ea5bc7cb30-kube-api-access-fnqj5\") pod \"313db7f6-e689-4071-87d8-71ea5bc7cb30\" (UID: \"313db7f6-e689-4071-87d8-71ea5bc7cb30\") " Oct 05 08:53:52 crc kubenswrapper[4935]: I1005 08:53:52.675353 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/313db7f6-e689-4071-87d8-71ea5bc7cb30-kube-api-access-fnqj5" (OuterVolumeSpecName: "kube-api-access-fnqj5") pod "313db7f6-e689-4071-87d8-71ea5bc7cb30" (UID: "313db7f6-e689-4071-87d8-71ea5bc7cb30"). InnerVolumeSpecName "kube-api-access-fnqj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:53:52 crc kubenswrapper[4935]: I1005 08:53:52.770123 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnqj5\" (UniqueName: \"kubernetes.io/projected/313db7f6-e689-4071-87d8-71ea5bc7cb30-kube-api-access-fnqj5\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:53 crc kubenswrapper[4935]: I1005 08:53:53.110652 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fd41-account-create-nkmgw" event={"ID":"313db7f6-e689-4071-87d8-71ea5bc7cb30","Type":"ContainerDied","Data":"94aef2e9c7d9e7eb14f95f9ea32feb6c6b176acb20c8b606afddb25f6c52c96c"} Oct 05 08:53:53 crc kubenswrapper[4935]: I1005 08:53:53.110691 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fd41-account-create-nkmgw" Oct 05 08:53:53 crc kubenswrapper[4935]: I1005 08:53:53.110712 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94aef2e9c7d9e7eb14f95f9ea32feb6c6b176acb20c8b606afddb25f6c52c96c" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.817514 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-kjp9m"] Oct 05 08:53:54 crc kubenswrapper[4935]: E1005 08:53:54.818081 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313db7f6-e689-4071-87d8-71ea5bc7cb30" containerName="mariadb-account-create" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.818098 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="313db7f6-e689-4071-87d8-71ea5bc7cb30" containerName="mariadb-account-create" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.818322 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="313db7f6-e689-4071-87d8-71ea5bc7cb30" containerName="mariadb-account-create" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.824285 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.831320 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.831622 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.831862 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jtxcd" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.838294 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kjp9m"] Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.922178 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-config-data\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.922249 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg2ms\" (UniqueName: \"kubernetes.io/projected/850ca766-4edf-4518-9e91-bf08ce1d119e-kube-api-access-lg2ms\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.922329 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-db-sync-config-data\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.922408 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-combined-ca-bundle\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.922614 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-scripts\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:54 crc kubenswrapper[4935]: I1005 08:53:54.922685 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/850ca766-4edf-4518-9e91-bf08ce1d119e-etc-machine-id\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.025599 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-db-sync-config-data\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.025687 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-combined-ca-bundle\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.025820 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-scripts\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.025868 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/850ca766-4edf-4518-9e91-bf08ce1d119e-etc-machine-id\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.025918 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-config-data\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.025957 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg2ms\" (UniqueName: \"kubernetes.io/projected/850ca766-4edf-4518-9e91-bf08ce1d119e-kube-api-access-lg2ms\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.026595 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/850ca766-4edf-4518-9e91-bf08ce1d119e-etc-machine-id\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.033617 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-db-sync-config-data\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.034199 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-scripts\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.034233 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-config-data\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.046798 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg2ms\" (UniqueName: \"kubernetes.io/projected/850ca766-4edf-4518-9e91-bf08ce1d119e-kube-api-access-lg2ms\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.050903 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-combined-ca-bundle\") pod \"cinder-db-sync-kjp9m\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.153984 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:53:55 crc kubenswrapper[4935]: I1005 08:53:55.611805 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kjp9m"] Oct 05 08:53:56 crc kubenswrapper[4935]: I1005 08:53:56.141040 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kjp9m" event={"ID":"850ca766-4edf-4518-9e91-bf08ce1d119e","Type":"ContainerStarted","Data":"dedbe5d007c2a39359a9a01a9cca031f20b458c5612534905387e570445fade5"} Oct 05 08:53:57 crc kubenswrapper[4935]: I1005 08:53:57.778488 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:53:57 crc kubenswrapper[4935]: E1005 08:53:57.778986 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:54:11 crc kubenswrapper[4935]: I1005 08:54:11.778083 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:54:11 crc kubenswrapper[4935]: E1005 08:54:11.778755 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 08:54:16 crc kubenswrapper[4935]: I1005 08:54:16.342705 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kjp9m" event={"ID":"850ca766-4edf-4518-9e91-bf08ce1d119e","Type":"ContainerStarted","Data":"d63098e78c0bb39eebbde224ac01bbefa97d873c7fda7d06b3a6e26adb2978d6"} Oct 05 08:54:16 crc kubenswrapper[4935]: I1005 08:54:16.362163 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-kjp9m" podStartSLOduration=2.953358224 podStartE2EDuration="22.362144984s" podCreationTimestamp="2025-10-05 08:53:54 +0000 UTC" firstStartedPulling="2025-10-05 08:53:55.615740042 +0000 UTC m=+7269.498366502" lastFinishedPulling="2025-10-05 08:54:15.024526802 +0000 UTC m=+7288.907153262" observedRunningTime="2025-10-05 08:54:16.355374914 +0000 UTC m=+7290.238001374" watchObservedRunningTime="2025-10-05 08:54:16.362144984 +0000 UTC m=+7290.244771444" Oct 05 08:54:18 crc kubenswrapper[4935]: I1005 08:54:18.359399 4935 generic.go:334] "Generic (PLEG): container finished" podID="850ca766-4edf-4518-9e91-bf08ce1d119e" containerID="d63098e78c0bb39eebbde224ac01bbefa97d873c7fda7d06b3a6e26adb2978d6" exitCode=0 Oct 05 08:54:18 crc kubenswrapper[4935]: I1005 08:54:18.359486 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kjp9m" event={"ID":"850ca766-4edf-4518-9e91-bf08ce1d119e","Type":"ContainerDied","Data":"d63098e78c0bb39eebbde224ac01bbefa97d873c7fda7d06b3a6e26adb2978d6"} Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.732846 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.875708 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg2ms\" (UniqueName: \"kubernetes.io/projected/850ca766-4edf-4518-9e91-bf08ce1d119e-kube-api-access-lg2ms\") pod \"850ca766-4edf-4518-9e91-bf08ce1d119e\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.875802 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/850ca766-4edf-4518-9e91-bf08ce1d119e-etc-machine-id\") pod \"850ca766-4edf-4518-9e91-bf08ce1d119e\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.875850 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-scripts\") pod \"850ca766-4edf-4518-9e91-bf08ce1d119e\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.875883 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-combined-ca-bundle\") pod \"850ca766-4edf-4518-9e91-bf08ce1d119e\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.875960 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-config-data\") pod \"850ca766-4edf-4518-9e91-bf08ce1d119e\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.876025 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-db-sync-config-data\") pod \"850ca766-4edf-4518-9e91-bf08ce1d119e\" (UID: \"850ca766-4edf-4518-9e91-bf08ce1d119e\") " Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.876562 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/850ca766-4edf-4518-9e91-bf08ce1d119e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "850ca766-4edf-4518-9e91-bf08ce1d119e" (UID: "850ca766-4edf-4518-9e91-bf08ce1d119e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.882444 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/850ca766-4edf-4518-9e91-bf08ce1d119e-kube-api-access-lg2ms" (OuterVolumeSpecName: "kube-api-access-lg2ms") pod "850ca766-4edf-4518-9e91-bf08ce1d119e" (UID: "850ca766-4edf-4518-9e91-bf08ce1d119e"). InnerVolumeSpecName "kube-api-access-lg2ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.883419 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "850ca766-4edf-4518-9e91-bf08ce1d119e" (UID: "850ca766-4edf-4518-9e91-bf08ce1d119e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.883801 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-scripts" (OuterVolumeSpecName: "scripts") pod "850ca766-4edf-4518-9e91-bf08ce1d119e" (UID: "850ca766-4edf-4518-9e91-bf08ce1d119e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.916853 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "850ca766-4edf-4518-9e91-bf08ce1d119e" (UID: "850ca766-4edf-4518-9e91-bf08ce1d119e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.945053 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-config-data" (OuterVolumeSpecName: "config-data") pod "850ca766-4edf-4518-9e91-bf08ce1d119e" (UID: "850ca766-4edf-4518-9e91-bf08ce1d119e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.978328 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.978676 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.978881 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.981089 4935 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850ca766-4edf-4518-9e91-bf08ce1d119e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.981174 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg2ms\" (UniqueName: \"kubernetes.io/projected/850ca766-4edf-4518-9e91-bf08ce1d119e-kube-api-access-lg2ms\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:19 crc kubenswrapper[4935]: I1005 08:54:19.981253 4935 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/850ca766-4edf-4518-9e91-bf08ce1d119e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.378601 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kjp9m" event={"ID":"850ca766-4edf-4518-9e91-bf08ce1d119e","Type":"ContainerDied","Data":"dedbe5d007c2a39359a9a01a9cca031f20b458c5612534905387e570445fade5"} Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.378915 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dedbe5d007c2a39359a9a01a9cca031f20b458c5612534905387e570445fade5" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.378981 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kjp9m" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.683315 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b47d67949-vkqlg"] Oct 05 08:54:20 crc kubenswrapper[4935]: E1005 08:54:20.683717 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850ca766-4edf-4518-9e91-bf08ce1d119e" containerName="cinder-db-sync" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.683740 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="850ca766-4edf-4518-9e91-bf08ce1d119e" containerName="cinder-db-sync" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.684000 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="850ca766-4edf-4518-9e91-bf08ce1d119e" containerName="cinder-db-sync" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.688046 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.698045 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b47d67949-vkqlg"] Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.798732 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-dns-svc\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.798778 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-config\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.798811 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.799104 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tgml\" (UniqueName: \"kubernetes.io/projected/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-kube-api-access-4tgml\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.799254 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.883884 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.889980 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.893581 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jtxcd" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.893842 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.894004 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.894856 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.905166 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.916170 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-dns-svc\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.916395 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-config\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.907248 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.916518 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.916956 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tgml\" (UniqueName: \"kubernetes.io/projected/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-kube-api-access-4tgml\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.917441 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.907025 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.918318 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-config\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.919135 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-dns-svc\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:20 crc kubenswrapper[4935]: I1005 08:54:20.940951 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tgml\" (UniqueName: \"kubernetes.io/projected/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-kube-api-access-4tgml\") pod \"dnsmasq-dns-7b47d67949-vkqlg\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.018999 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.019348 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-scripts\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.019494 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8fd19384-485b-4224-ae32-5f54319214d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.019524 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fd19384-485b-4224-ae32-5f54319214d4-logs\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.019555 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv5s5\" (UniqueName: \"kubernetes.io/projected/8fd19384-485b-4224-ae32-5f54319214d4-kube-api-access-jv5s5\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.019591 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.019868 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.019985 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.121423 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-scripts\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.121507 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8fd19384-485b-4224-ae32-5f54319214d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.121539 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fd19384-485b-4224-ae32-5f54319214d4-logs\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.121571 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv5s5\" (UniqueName: \"kubernetes.io/projected/8fd19384-485b-4224-ae32-5f54319214d4-kube-api-access-jv5s5\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.121611 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.121645 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.121674 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.122573 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8fd19384-485b-4224-ae32-5f54319214d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.122942 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fd19384-485b-4224-ae32-5f54319214d4-logs\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.127747 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.134415 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.137815 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.140337 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-scripts\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.142174 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv5s5\" (UniqueName: \"kubernetes.io/projected/8fd19384-485b-4224-ae32-5f54319214d4-kube-api-access-jv5s5\") pod \"cinder-api-0\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.223244 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.489925 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b47d67949-vkqlg"] Oct 05 08:54:21 crc kubenswrapper[4935]: I1005 08:54:21.673002 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:54:22 crc kubenswrapper[4935]: I1005 08:54:22.402011 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8fd19384-485b-4224-ae32-5f54319214d4","Type":"ContainerStarted","Data":"1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb"} Oct 05 08:54:22 crc kubenswrapper[4935]: I1005 08:54:22.402303 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8fd19384-485b-4224-ae32-5f54319214d4","Type":"ContainerStarted","Data":"82c3ed077af29a72a8116b958bf438b7e2c84ab4c3d3c33aa01fc22a142cf4c9"} Oct 05 08:54:22 crc kubenswrapper[4935]: I1005 08:54:22.405157 4935 generic.go:334] "Generic (PLEG): container finished" podID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" containerID="77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320" exitCode=0 Oct 05 08:54:22 crc kubenswrapper[4935]: I1005 08:54:22.405195 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" event={"ID":"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f","Type":"ContainerDied","Data":"77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320"} Oct 05 08:54:22 crc kubenswrapper[4935]: I1005 08:54:22.405228 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" event={"ID":"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f","Type":"ContainerStarted","Data":"4a84f78519c144f76ee595e0be1bc05d3f6bb5802c1bb52960dce02971cd03e4"} Oct 05 08:54:23 crc kubenswrapper[4935]: I1005 08:54:23.418572 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" event={"ID":"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f","Type":"ContainerStarted","Data":"1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa"} Oct 05 08:54:23 crc kubenswrapper[4935]: I1005 08:54:23.418815 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:23 crc kubenswrapper[4935]: I1005 08:54:23.422511 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8fd19384-485b-4224-ae32-5f54319214d4","Type":"ContainerStarted","Data":"a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f"} Oct 05 08:54:23 crc kubenswrapper[4935]: I1005 08:54:23.422656 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 08:54:23 crc kubenswrapper[4935]: I1005 08:54:23.455326 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" podStartSLOduration=3.4553037939999998 podStartE2EDuration="3.455303794s" podCreationTimestamp="2025-10-05 08:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:54:23.441371665 +0000 UTC m=+7297.323998165" watchObservedRunningTime="2025-10-05 08:54:23.455303794 +0000 UTC m=+7297.337930254" Oct 05 08:54:23 crc kubenswrapper[4935]: I1005 08:54:23.468759 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.468737831 podStartE2EDuration="3.468737831s" podCreationTimestamp="2025-10-05 08:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:54:23.465749912 +0000 UTC m=+7297.348376372" watchObservedRunningTime="2025-10-05 08:54:23.468737831 +0000 UTC m=+7297.351364291" Oct 05 08:54:24 crc kubenswrapper[4935]: I1005 08:54:24.777986 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:54:25 crc kubenswrapper[4935]: I1005 08:54:25.450994 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"4f78b54294a258ac737523e93c508737226a48f0df445e1a69f2784b25cc51ab"} Oct 05 08:54:31 crc kubenswrapper[4935]: I1005 08:54:31.020082 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 08:54:31 crc kubenswrapper[4935]: I1005 08:54:31.109570 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648568c79-jw292"] Oct 05 08:54:31 crc kubenswrapper[4935]: I1005 08:54:31.110457 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-648568c79-jw292" podUID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" containerName="dnsmasq-dns" containerID="cri-o://058b4fb7475b58ff5899f48eaa3828f618db23c42f22f67cd913e50c3b6f87d2" gracePeriod=10 Oct 05 08:54:31 crc kubenswrapper[4935]: I1005 08:54:31.528649 4935 generic.go:334] "Generic (PLEG): container finished" podID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" containerID="058b4fb7475b58ff5899f48eaa3828f618db23c42f22f67cd913e50c3b6f87d2" exitCode=0 Oct 05 08:54:31 crc kubenswrapper[4935]: I1005 08:54:31.528692 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648568c79-jw292" event={"ID":"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d","Type":"ContainerDied","Data":"058b4fb7475b58ff5899f48eaa3828f618db23c42f22f67cd913e50c3b6f87d2"} Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.068657 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.234439 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-nb\") pod \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.234563 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-sb\") pod \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.234614 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-dns-svc\") pod \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.234673 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-config\") pod \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.234755 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7slr8\" (UniqueName: \"kubernetes.io/projected/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-kube-api-access-7slr8\") pod \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\" (UID: \"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d\") " Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.241882 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-kube-api-access-7slr8" (OuterVolumeSpecName: "kube-api-access-7slr8") pod "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" (UID: "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d"). InnerVolumeSpecName "kube-api-access-7slr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.293513 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" (UID: "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.294642 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-config" (OuterVolumeSpecName: "config") pod "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" (UID: "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.327866 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" (UID: "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.336976 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.337003 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.337017 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.337031 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7slr8\" (UniqueName: \"kubernetes.io/projected/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-kube-api-access-7slr8\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.341385 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" (UID: "7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.438682 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.541373 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648568c79-jw292" event={"ID":"7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d","Type":"ContainerDied","Data":"310f2c1409d80af8c81ad8f8a78f6533a7831b8cb2eb140683e829acf9857407"} Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.541430 4935 scope.go:117] "RemoveContainer" containerID="058b4fb7475b58ff5899f48eaa3828f618db23c42f22f67cd913e50c3b6f87d2" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.541488 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648568c79-jw292" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.571477 4935 scope.go:117] "RemoveContainer" containerID="7fd479a665b838529e89e194e7b197f88ec85fa155ef4597f8094eddd230e14f" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.574459 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648568c79-jw292"] Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.582574 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-648568c79-jw292"] Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.787106 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" path="/var/lib/kubelet/pods/7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d/volumes" Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.844034 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.844536 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-log" containerID="cri-o://778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3" gracePeriod=30 Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.844658 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-metadata" containerID="cri-o://28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e" gracePeriod=30 Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.858788 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.859082 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0cd2e2a0-6247-45b3-a8c3-df740829908f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://4879cd0b51ae4570f9468ff413c3752a280ad7596a1ad09ecfa714142128d33e" gracePeriod=30 Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.884383 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.884708 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.884931 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="76e7a7c0-e906-4e2d-95e1-0062edf3ff36" containerName="nova-cell0-conductor-conductor" containerID="cri-o://af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1" gracePeriod=30 Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.885174 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="49137eff-8dd5-4896-9770-77a953079e83" containerName="nova-scheduler-scheduler" containerID="cri-o://82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87" gracePeriod=30 Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.912998 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.913564 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-log" containerID="cri-o://fe12d4a3ada614e73e199d1a7d8f93d5fb49ef898df4b349402244188a337eb9" gracePeriod=30 Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.914032 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-api" containerID="cri-o://5784c136d0f06f0ed1fa0e7636182c1d63a5241795a6a516c6f12baf39969bd5" gracePeriod=30 Oct 05 08:54:32 crc kubenswrapper[4935]: I1005 08:54:32.966869 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.556347 4935 generic.go:334] "Generic (PLEG): container finished" podID="2f473afb-6b80-4c81-83e3-140d713f354f" containerID="fe12d4a3ada614e73e199d1a7d8f93d5fb49ef898df4b349402244188a337eb9" exitCode=143 Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.556451 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f473afb-6b80-4c81-83e3-140d713f354f","Type":"ContainerDied","Data":"fe12d4a3ada614e73e199d1a7d8f93d5fb49ef898df4b349402244188a337eb9"} Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.558468 4935 generic.go:334] "Generic (PLEG): container finished" podID="0cd2e2a0-6247-45b3-a8c3-df740829908f" containerID="4879cd0b51ae4570f9468ff413c3752a280ad7596a1ad09ecfa714142128d33e" exitCode=0 Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.558513 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0cd2e2a0-6247-45b3-a8c3-df740829908f","Type":"ContainerDied","Data":"4879cd0b51ae4570f9468ff413c3752a280ad7596a1ad09ecfa714142128d33e"} Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.564111 4935 generic.go:334] "Generic (PLEG): container finished" podID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerID="778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3" exitCode=143 Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.564173 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1","Type":"ContainerDied","Data":"778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3"} Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.790844 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.872457 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-config-data\") pod \"0cd2e2a0-6247-45b3-a8c3-df740829908f\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.872591 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmtzj\" (UniqueName: \"kubernetes.io/projected/0cd2e2a0-6247-45b3-a8c3-df740829908f-kube-api-access-qmtzj\") pod \"0cd2e2a0-6247-45b3-a8c3-df740829908f\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.872740 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-combined-ca-bundle\") pod \"0cd2e2a0-6247-45b3-a8c3-df740829908f\" (UID: \"0cd2e2a0-6247-45b3-a8c3-df740829908f\") " Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.893173 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cd2e2a0-6247-45b3-a8c3-df740829908f-kube-api-access-qmtzj" (OuterVolumeSpecName: "kube-api-access-qmtzj") pod "0cd2e2a0-6247-45b3-a8c3-df740829908f" (UID: "0cd2e2a0-6247-45b3-a8c3-df740829908f"). InnerVolumeSpecName "kube-api-access-qmtzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.924594 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cd2e2a0-6247-45b3-a8c3-df740829908f" (UID: "0cd2e2a0-6247-45b3-a8c3-df740829908f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.928354 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-config-data" (OuterVolumeSpecName: "config-data") pod "0cd2e2a0-6247-45b3-a8c3-df740829908f" (UID: "0cd2e2a0-6247-45b3-a8c3-df740829908f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.975896 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.975951 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmtzj\" (UniqueName: \"kubernetes.io/projected/0cd2e2a0-6247-45b3-a8c3-df740829908f-kube-api-access-qmtzj\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.975965 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd2e2a0-6247-45b3-a8c3-df740829908f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:33 crc kubenswrapper[4935]: I1005 08:54:33.990873 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.076898 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-combined-ca-bundle\") pod \"49137eff-8dd5-4896-9770-77a953079e83\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.077057 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-config-data\") pod \"49137eff-8dd5-4896-9770-77a953079e83\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.077123 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q87x8\" (UniqueName: \"kubernetes.io/projected/49137eff-8dd5-4896-9770-77a953079e83-kube-api-access-q87x8\") pod \"49137eff-8dd5-4896-9770-77a953079e83\" (UID: \"49137eff-8dd5-4896-9770-77a953079e83\") " Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.081223 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49137eff-8dd5-4896-9770-77a953079e83-kube-api-access-q87x8" (OuterVolumeSpecName: "kube-api-access-q87x8") pod "49137eff-8dd5-4896-9770-77a953079e83" (UID: "49137eff-8dd5-4896-9770-77a953079e83"). InnerVolumeSpecName "kube-api-access-q87x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.099488 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-config-data" (OuterVolumeSpecName: "config-data") pod "49137eff-8dd5-4896-9770-77a953079e83" (UID: "49137eff-8dd5-4896-9770-77a953079e83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.104999 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49137eff-8dd5-4896-9770-77a953079e83" (UID: "49137eff-8dd5-4896-9770-77a953079e83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.179538 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.179577 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49137eff-8dd5-4896-9770-77a953079e83-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.179589 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q87x8\" (UniqueName: \"kubernetes.io/projected/49137eff-8dd5-4896-9770-77a953079e83-kube-api-access-q87x8\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.317345 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.320428 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.321887 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.321933 4935 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="76e7a7c0-e906-4e2d-95e1-0062edf3ff36" containerName="nova-cell0-conductor-conductor" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.578611 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.581813 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0cd2e2a0-6247-45b3-a8c3-df740829908f","Type":"ContainerDied","Data":"959b9143bcd44010cca5b076ea126afd4305aecf3882f706b10fbbcc88bc19a2"} Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.581884 4935 scope.go:117] "RemoveContainer" containerID="4879cd0b51ae4570f9468ff413c3752a280ad7596a1ad09ecfa714142128d33e" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.584798 4935 generic.go:334] "Generic (PLEG): container finished" podID="49137eff-8dd5-4896-9770-77a953079e83" containerID="82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87" exitCode=0 Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.584842 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"49137eff-8dd5-4896-9770-77a953079e83","Type":"ContainerDied","Data":"82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87"} Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.584870 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"49137eff-8dd5-4896-9770-77a953079e83","Type":"ContainerDied","Data":"280504ca9e05db6128fa92fb6187e0e0ebe929882ce918221a6b5cdb501058ee"} Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.585087 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.631655 4935 scope.go:117] "RemoveContainer" containerID="82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.642381 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.651951 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.660265 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.671582 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.672152 4935 scope.go:117] "RemoveContainer" containerID="82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87" Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.675353 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87\": container with ID starting with 82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87 not found: ID does not exist" containerID="82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.675470 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87"} err="failed to get container status \"82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87\": rpc error: code = NotFound desc = could not find container \"82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87\": container with ID starting with 82eff527bfb99a4e4c7dd8935f76f71d69db70c044b47985e0382e901695cf87 not found: ID does not exist" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.685199 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.685748 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" containerName="dnsmasq-dns" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.685769 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" containerName="dnsmasq-dns" Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.685804 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" containerName="init" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.685813 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" containerName="init" Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.685832 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd2e2a0-6247-45b3-a8c3-df740829908f" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.685841 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd2e2a0-6247-45b3-a8c3-df740829908f" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 08:54:34 crc kubenswrapper[4935]: E1005 08:54:34.685997 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49137eff-8dd5-4896-9770-77a953079e83" containerName="nova-scheduler-scheduler" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.686008 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="49137eff-8dd5-4896-9770-77a953079e83" containerName="nova-scheduler-scheduler" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.686263 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cd2e2a0-6247-45b3-a8c3-df740829908f" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.686288 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a8dcc46-ccc4-4ac0-be8a-4bf4bd5edc4d" containerName="dnsmasq-dns" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.686305 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="49137eff-8dd5-4896-9770-77a953079e83" containerName="nova-scheduler-scheduler" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.693313 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.700699 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.703012 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.718109 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.722327 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.724579 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.727082 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.795192 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cd2e2a0-6247-45b3-a8c3-df740829908f" path="/var/lib/kubelet/pods/0cd2e2a0-6247-45b3-a8c3-df740829908f/volumes" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.795900 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49137eff-8dd5-4896-9770-77a953079e83" path="/var/lib/kubelet/pods/49137eff-8dd5-4896-9770-77a953079e83/volumes" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.807593 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvd2k\" (UniqueName: \"kubernetes.io/projected/d5ff7e17-3100-4f92-a63d-fdd61864a272-kube-api-access-vvd2k\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.807708 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.807817 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-config-data\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.909830 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.909955 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.910010 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-config-data\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.910141 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvd2k\" (UniqueName: \"kubernetes.io/projected/d5ff7e17-3100-4f92-a63d-fdd61864a272-kube-api-access-vvd2k\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.910183 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.910208 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9q5b\" (UniqueName: \"kubernetes.io/projected/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-kube-api-access-b9q5b\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.920286 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-config-data\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.925518 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:34 crc kubenswrapper[4935]: I1005 08:54:34.931239 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvd2k\" (UniqueName: \"kubernetes.io/projected/d5ff7e17-3100-4f92-a63d-fdd61864a272-kube-api-access-vvd2k\") pod \"nova-scheduler-0\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " pod="openstack/nova-scheduler-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.011738 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.011785 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9q5b\" (UniqueName: \"kubernetes.io/projected/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-kube-api-access-b9q5b\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.011839 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.019711 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.023006 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.025611 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.043839 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9q5b\" (UniqueName: \"kubernetes.io/projected/7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1-kube-api-access-b9q5b\") pod \"nova-cell1-novncproxy-0\" (UID: \"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.342599 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.564244 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.597072 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d5ff7e17-3100-4f92-a63d-fdd61864a272","Type":"ContainerStarted","Data":"2755422a86994a32e0eee1a1ac2e4314171ac81769b54ecc04af30f7812f8ef7"} Oct 05 08:54:35 crc kubenswrapper[4935]: I1005 08:54:35.783509 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.052873 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": read tcp 10.217.0.2:35932->10.217.1.84:8775: read: connection reset by peer" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.053352 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": read tcp 10.217.0.2:35920->10.217.1.84:8775: read: connection reset by peer" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.287814 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.85:8774/\": read tcp 10.217.0.2:42288->10.217.1.85:8774: read: connection reset by peer" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.287961 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.85:8774/\": read tcp 10.217.0.2:42276->10.217.1.85:8774: read: connection reset by peer" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.337783 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.338048 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="2d0f44d4-a284-4c21-bea7-b964477081c4" containerName="nova-cell1-conductor-conductor" containerID="cri-o://d816192d2da7464a7d00b60d0018a5bea686b0f90a2a1722ee561a918945d8d7" gracePeriod=30 Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.591107 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.626154 4935 generic.go:334] "Generic (PLEG): container finished" podID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerID="28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e" exitCode=0 Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.626336 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.627234 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1","Type":"ContainerDied","Data":"28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e"} Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.627265 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1","Type":"ContainerDied","Data":"9e4e0e430cd7be262998c776883ae0e2a5807572e21e394c6812e78703b2f431"} Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.627282 4935 scope.go:117] "RemoveContainer" containerID="28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.645260 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d5ff7e17-3100-4f92-a63d-fdd61864a272","Type":"ContainerStarted","Data":"446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28"} Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.652215 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1","Type":"ContainerStarted","Data":"3fcce55a3de943124220d8fdffcac9f206e73b33bf7c3a3c72bd9a4da3ade9ce"} Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.652266 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1","Type":"ContainerStarted","Data":"58b90bc180e83b766bbe499a522816a8a8fa9eac24fba2da51e8121a6a97eb2e"} Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.662323 4935 generic.go:334] "Generic (PLEG): container finished" podID="2f473afb-6b80-4c81-83e3-140d713f354f" containerID="5784c136d0f06f0ed1fa0e7636182c1d63a5241795a6a516c6f12baf39969bd5" exitCode=0 Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.662360 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f473afb-6b80-4c81-83e3-140d713f354f","Type":"ContainerDied","Data":"5784c136d0f06f0ed1fa0e7636182c1d63a5241795a6a516c6f12baf39969bd5"} Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.678242 4935 scope.go:117] "RemoveContainer" containerID="778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.704622 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.704600679 podStartE2EDuration="2.704600679s" podCreationTimestamp="2025-10-05 08:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:54:36.697752858 +0000 UTC m=+7310.580379318" watchObservedRunningTime="2025-10-05 08:54:36.704600679 +0000 UTC m=+7310.587227139" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.705450 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.713835 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.713809323 podStartE2EDuration="2.713809323s" podCreationTimestamp="2025-10-05 08:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:54:36.670623278 +0000 UTC m=+7310.553249738" watchObservedRunningTime="2025-10-05 08:54:36.713809323 +0000 UTC m=+7310.596435783" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.724687 4935 scope.go:117] "RemoveContainer" containerID="28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e" Oct 05 08:54:36 crc kubenswrapper[4935]: E1005 08:54:36.731142 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e\": container with ID starting with 28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e not found: ID does not exist" containerID="28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.731192 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e"} err="failed to get container status \"28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e\": rpc error: code = NotFound desc = could not find container \"28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e\": container with ID starting with 28d49b1ba47a733bd3bf52d8194adb414a36e50c173d73fb8bd9c3072f49064e not found: ID does not exist" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.731215 4935 scope.go:117] "RemoveContainer" containerID="778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3" Oct 05 08:54:36 crc kubenswrapper[4935]: E1005 08:54:36.732372 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3\": container with ID starting with 778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3 not found: ID does not exist" containerID="778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.732392 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3"} err="failed to get container status \"778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3\": rpc error: code = NotFound desc = could not find container \"778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3\": container with ID starting with 778733a5eff0ce161cc1ae89f2904542751dbff38dd6c8bf3d7006c6782041a3 not found: ID does not exist" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.746553 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-config-data\") pod \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.746629 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-combined-ca-bundle\") pod \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.746747 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-logs\") pod \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.746896 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fztsr\" (UniqueName: \"kubernetes.io/projected/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-kube-api-access-fztsr\") pod \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\" (UID: \"6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1\") " Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.752298 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-logs" (OuterVolumeSpecName: "logs") pod "6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" (UID: "6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.764225 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-kube-api-access-fztsr" (OuterVolumeSpecName: "kube-api-access-fztsr") pod "6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" (UID: "6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1"). InnerVolumeSpecName "kube-api-access-fztsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.779235 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" (UID: "6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.785079 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-config-data" (OuterVolumeSpecName: "config-data") pod "6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" (UID: "6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.849154 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwtx9\" (UniqueName: \"kubernetes.io/projected/2f473afb-6b80-4c81-83e3-140d713f354f-kube-api-access-vwtx9\") pod \"2f473afb-6b80-4c81-83e3-140d713f354f\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.849734 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-config-data\") pod \"2f473afb-6b80-4c81-83e3-140d713f354f\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.849803 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f473afb-6b80-4c81-83e3-140d713f354f-logs\") pod \"2f473afb-6b80-4c81-83e3-140d713f354f\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.849969 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-combined-ca-bundle\") pod \"2f473afb-6b80-4c81-83e3-140d713f354f\" (UID: \"2f473afb-6b80-4c81-83e3-140d713f354f\") " Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.850569 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.850595 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fztsr\" (UniqueName: \"kubernetes.io/projected/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-kube-api-access-fztsr\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.850609 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.850621 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.850949 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f473afb-6b80-4c81-83e3-140d713f354f-logs" (OuterVolumeSpecName: "logs") pod "2f473afb-6b80-4c81-83e3-140d713f354f" (UID: "2f473afb-6b80-4c81-83e3-140d713f354f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.852153 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f473afb-6b80-4c81-83e3-140d713f354f-kube-api-access-vwtx9" (OuterVolumeSpecName: "kube-api-access-vwtx9") pod "2f473afb-6b80-4c81-83e3-140d713f354f" (UID: "2f473afb-6b80-4c81-83e3-140d713f354f"). InnerVolumeSpecName "kube-api-access-vwtx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.883198 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f473afb-6b80-4c81-83e3-140d713f354f" (UID: "2f473afb-6b80-4c81-83e3-140d713f354f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.884127 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-config-data" (OuterVolumeSpecName: "config-data") pod "2f473afb-6b80-4c81-83e3-140d713f354f" (UID: "2f473afb-6b80-4c81-83e3-140d713f354f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.953046 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.953111 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f473afb-6b80-4c81-83e3-140d713f354f-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.953126 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f473afb-6b80-4c81-83e3-140d713f354f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.953146 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwtx9\" (UniqueName: \"kubernetes.io/projected/2f473afb-6b80-4c81-83e3-140d713f354f-kube-api-access-vwtx9\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.960711 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.972373 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.981178 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:54:36 crc kubenswrapper[4935]: E1005 08:54:36.981767 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-log" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.981789 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-log" Oct 05 08:54:36 crc kubenswrapper[4935]: E1005 08:54:36.981812 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-log" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.981820 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-log" Oct 05 08:54:36 crc kubenswrapper[4935]: E1005 08:54:36.981848 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-metadata" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.981854 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-metadata" Oct 05 08:54:36 crc kubenswrapper[4935]: E1005 08:54:36.981897 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-api" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.981903 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-api" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.982157 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-log" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.982174 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" containerName="nova-api-api" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.982188 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-metadata" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.982196 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" containerName="nova-metadata-log" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.983584 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:54:36 crc kubenswrapper[4935]: I1005 08:54:36.997298 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.015296 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.157368 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blf55\" (UniqueName: \"kubernetes.io/projected/bc0befef-da68-4fa7-b273-e030be230f3b-kube-api-access-blf55\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.157524 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.157561 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-config-data\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.157828 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc0befef-da68-4fa7-b273-e030be230f3b-logs\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.260128 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.260241 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-config-data\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.260403 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc0befef-da68-4fa7-b273-e030be230f3b-logs\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.260466 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blf55\" (UniqueName: \"kubernetes.io/projected/bc0befef-da68-4fa7-b273-e030be230f3b-kube-api-access-blf55\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.260867 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc0befef-da68-4fa7-b273-e030be230f3b-logs\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.264094 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.265843 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-config-data\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.278387 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blf55\" (UniqueName: \"kubernetes.io/projected/bc0befef-da68-4fa7-b273-e030be230f3b-kube-api-access-blf55\") pod \"nova-metadata-0\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.321667 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.398670 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v5wdx"] Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.400978 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.407794 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5wdx"] Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.565163 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-utilities\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.565543 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w4p6\" (UniqueName: \"kubernetes.io/projected/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-kube-api-access-7w4p6\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.565586 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-catalog-content\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.669242 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-utilities\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.669332 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w4p6\" (UniqueName: \"kubernetes.io/projected/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-kube-api-access-7w4p6\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.669367 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-catalog-content\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.670080 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-catalog-content\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.674339 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-utilities\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.684369 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.685010 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f473afb-6b80-4c81-83e3-140d713f354f","Type":"ContainerDied","Data":"4a7a28e72b7aee223e0d31f407b2416e2db85e3b7564f74358ce3b66c565ff60"} Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.685280 4935 scope.go:117] "RemoveContainer" containerID="5784c136d0f06f0ed1fa0e7636182c1d63a5241795a6a516c6f12baf39969bd5" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.699165 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w4p6\" (UniqueName: \"kubernetes.io/projected/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-kube-api-access-7w4p6\") pod \"certified-operators-v5wdx\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.768323 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.771804 4935 scope.go:117] "RemoveContainer" containerID="fe12d4a3ada614e73e199d1a7d8f93d5fb49ef898df4b349402244188a337eb9" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.777027 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.813166 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.840731 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.842522 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.847284 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.850215 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:54:37 crc kubenswrapper[4935]: W1005 08:54:37.975085 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc0befef_da68_4fa7_b273_e030be230f3b.slice/crio-05599728c663204ab9d7f2bc3a07a6c97a11623a0d83265ba865b43ac30f3e58 WatchSource:0}: Error finding container 05599728c663204ab9d7f2bc3a07a6c97a11623a0d83265ba865b43ac30f3e58: Status 404 returned error can't find the container with id 05599728c663204ab9d7f2bc3a07a6c97a11623a0d83265ba865b43ac30f3e58 Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.981674 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d066b6e6-5017-42da-9868-0d61d229f199-logs\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.981729 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-config-data\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.981953 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.981993 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p86wl\" (UniqueName: \"kubernetes.io/projected/d066b6e6-5017-42da-9868-0d61d229f199-kube-api-access-p86wl\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:37 crc kubenswrapper[4935]: I1005 08:54:37.991993 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.109130 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.109178 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p86wl\" (UniqueName: \"kubernetes.io/projected/d066b6e6-5017-42da-9868-0d61d229f199-kube-api-access-p86wl\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.109226 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d066b6e6-5017-42da-9868-0d61d229f199-logs\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.109245 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-config-data\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.110677 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d066b6e6-5017-42da-9868-0d61d229f199-logs\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.118992 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.120249 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-config-data\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.131384 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p86wl\" (UniqueName: \"kubernetes.io/projected/d066b6e6-5017-42da-9868-0d61d229f199-kube-api-access-p86wl\") pod \"nova-api-0\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.184173 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.412545 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5wdx"] Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.515168 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:54:38 crc kubenswrapper[4935]: W1005 08:54:38.533296 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd066b6e6_5017_42da_9868_0d61d229f199.slice/crio-be36e01b91f6ae31c3dd4f6162e509a2da661bf0725d08de3b534d33228bbe54 WatchSource:0}: Error finding container be36e01b91f6ae31c3dd4f6162e509a2da661bf0725d08de3b534d33228bbe54: Status 404 returned error can't find the container with id be36e01b91f6ae31c3dd4f6162e509a2da661bf0725d08de3b534d33228bbe54 Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.712821 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5wdx" event={"ID":"a8775ed9-5817-4c33-87e6-a1119d4ccc7b","Type":"ContainerStarted","Data":"dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.712860 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5wdx" event={"ID":"a8775ed9-5817-4c33-87e6-a1119d4ccc7b","Type":"ContainerStarted","Data":"4e5b4d196dbefc7997378a45c7e9545e8f17886ebbf418464d903e2973cc06dd"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.724200 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d066b6e6-5017-42da-9868-0d61d229f199","Type":"ContainerStarted","Data":"be36e01b91f6ae31c3dd4f6162e509a2da661bf0725d08de3b534d33228bbe54"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.741479 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc0befef-da68-4fa7-b273-e030be230f3b","Type":"ContainerStarted","Data":"212013d09ca844d3947df7c252d6509ae573f77629ee77ac4068ce10aa8e733b"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.741529 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc0befef-da68-4fa7-b273-e030be230f3b","Type":"ContainerStarted","Data":"27be07cf923f7e51489ec40bf7111a7421289df206b9848709b133d828374621"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.741541 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc0befef-da68-4fa7-b273-e030be230f3b","Type":"ContainerStarted","Data":"05599728c663204ab9d7f2bc3a07a6c97a11623a0d83265ba865b43ac30f3e58"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.744105 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.744778 4935 generic.go:334] "Generic (PLEG): container finished" podID="76e7a7c0-e906-4e2d-95e1-0062edf3ff36" containerID="af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1" exitCode=0 Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.744834 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"76e7a7c0-e906-4e2d-95e1-0062edf3ff36","Type":"ContainerDied","Data":"af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.744858 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"76e7a7c0-e906-4e2d-95e1-0062edf3ff36","Type":"ContainerDied","Data":"b49133c1bbce63709e7bafd77cd07148f9fa7d50aa3afce7e9411ef417cb5bc3"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.744876 4935 scope.go:117] "RemoveContainer" containerID="af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.750595 4935 generic.go:334] "Generic (PLEG): container finished" podID="2d0f44d4-a284-4c21-bea7-b964477081c4" containerID="d816192d2da7464a7d00b60d0018a5bea686b0f90a2a1722ee561a918945d8d7" exitCode=0 Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.750647 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2d0f44d4-a284-4c21-bea7-b964477081c4","Type":"ContainerDied","Data":"d816192d2da7464a7d00b60d0018a5bea686b0f90a2a1722ee561a918945d8d7"} Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.763274 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7632535000000003 podStartE2EDuration="2.7632535s" podCreationTimestamp="2025-10-05 08:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:54:38.759815639 +0000 UTC m=+7312.642442109" watchObservedRunningTime="2025-10-05 08:54:38.7632535 +0000 UTC m=+7312.645879960" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.781972 4935 scope.go:117] "RemoveContainer" containerID="af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1" Oct 05 08:54:38 crc kubenswrapper[4935]: E1005 08:54:38.789007 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1\": container with ID starting with af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1 not found: ID does not exist" containerID="af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.789139 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1"} err="failed to get container status \"af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1\": rpc error: code = NotFound desc = could not find container \"af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1\": container with ID starting with af5b5e95d983113b1a4f8a3d4256a1da549e7dff730e0ab2c0791f100b1480b1 not found: ID does not exist" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.797983 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f473afb-6b80-4c81-83e3-140d713f354f" path="/var/lib/kubelet/pods/2f473afb-6b80-4c81-83e3-140d713f354f/volumes" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.798786 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1" path="/var/lib/kubelet/pods/6069ef15-6a67-4ba8-9dc2-9a5ecc8adfb1/volumes" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.827861 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-config-data\") pod \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.828111 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-combined-ca-bundle\") pod \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.828237 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqnp7\" (UniqueName: \"kubernetes.io/projected/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-kube-api-access-sqnp7\") pod \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\" (UID: \"76e7a7c0-e906-4e2d-95e1-0062edf3ff36\") " Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.834464 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-kube-api-access-sqnp7" (OuterVolumeSpecName: "kube-api-access-sqnp7") pod "76e7a7c0-e906-4e2d-95e1-0062edf3ff36" (UID: "76e7a7c0-e906-4e2d-95e1-0062edf3ff36"). InnerVolumeSpecName "kube-api-access-sqnp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.871077 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-config-data" (OuterVolumeSpecName: "config-data") pod "76e7a7c0-e906-4e2d-95e1-0062edf3ff36" (UID: "76e7a7c0-e906-4e2d-95e1-0062edf3ff36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.877064 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76e7a7c0-e906-4e2d-95e1-0062edf3ff36" (UID: "76e7a7c0-e906-4e2d-95e1-0062edf3ff36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.929875 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.929922 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.929933 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqnp7\" (UniqueName: \"kubernetes.io/projected/76e7a7c0-e906-4e2d-95e1-0062edf3ff36-kube-api-access-sqnp7\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:38 crc kubenswrapper[4935]: I1005 08:54:38.971383 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.133330 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d76v6\" (UniqueName: \"kubernetes.io/projected/2d0f44d4-a284-4c21-bea7-b964477081c4-kube-api-access-d76v6\") pod \"2d0f44d4-a284-4c21-bea7-b964477081c4\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.133579 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-combined-ca-bundle\") pod \"2d0f44d4-a284-4c21-bea7-b964477081c4\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.133627 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-config-data\") pod \"2d0f44d4-a284-4c21-bea7-b964477081c4\" (UID: \"2d0f44d4-a284-4c21-bea7-b964477081c4\") " Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.136739 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d0f44d4-a284-4c21-bea7-b964477081c4-kube-api-access-d76v6" (OuterVolumeSpecName: "kube-api-access-d76v6") pod "2d0f44d4-a284-4c21-bea7-b964477081c4" (UID: "2d0f44d4-a284-4c21-bea7-b964477081c4"). InnerVolumeSpecName "kube-api-access-d76v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.162459 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-config-data" (OuterVolumeSpecName: "config-data") pod "2d0f44d4-a284-4c21-bea7-b964477081c4" (UID: "2d0f44d4-a284-4c21-bea7-b964477081c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.173016 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d0f44d4-a284-4c21-bea7-b964477081c4" (UID: "2d0f44d4-a284-4c21-bea7-b964477081c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.235401 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.235433 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0f44d4-a284-4c21-bea7-b964477081c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.235444 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d76v6\" (UniqueName: \"kubernetes.io/projected/2d0f44d4-a284-4c21-bea7-b964477081c4-kube-api-access-d76v6\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.759894 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.762984 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2d0f44d4-a284-4c21-bea7-b964477081c4","Type":"ContainerDied","Data":"d620ca9c46e083e75547ff906b742910a81e49c2215470687f193745cdf30b4a"} Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.763043 4935 scope.go:117] "RemoveContainer" containerID="d816192d2da7464a7d00b60d0018a5bea686b0f90a2a1722ee561a918945d8d7" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.763148 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.774266 4935 generic.go:334] "Generic (PLEG): container finished" podID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerID="dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f" exitCode=0 Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.774336 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5wdx" event={"ID":"a8775ed9-5817-4c33-87e6-a1119d4ccc7b","Type":"ContainerDied","Data":"dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f"} Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.774362 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5wdx" event={"ID":"a8775ed9-5817-4c33-87e6-a1119d4ccc7b","Type":"ContainerStarted","Data":"354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2"} Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.777705 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d066b6e6-5017-42da-9868-0d61d229f199","Type":"ContainerStarted","Data":"8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d"} Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.777736 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d066b6e6-5017-42da-9868-0d61d229f199","Type":"ContainerStarted","Data":"9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf"} Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.819347 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.840117 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.856711 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.857038 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.857020928 podStartE2EDuration="2.857020928s" podCreationTimestamp="2025-10-05 08:54:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:54:39.834308976 +0000 UTC m=+7313.716935436" watchObservedRunningTime="2025-10-05 08:54:39.857020928 +0000 UTC m=+7313.739647388" Oct 05 08:54:39 crc kubenswrapper[4935]: E1005 08:54:39.857188 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d0f44d4-a284-4c21-bea7-b964477081c4" containerName="nova-cell1-conductor-conductor" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.857212 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d0f44d4-a284-4c21-bea7-b964477081c4" containerName="nova-cell1-conductor-conductor" Oct 05 08:54:39 crc kubenswrapper[4935]: E1005 08:54:39.857251 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e7a7c0-e906-4e2d-95e1-0062edf3ff36" containerName="nova-cell0-conductor-conductor" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.857261 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e7a7c0-e906-4e2d-95e1-0062edf3ff36" containerName="nova-cell0-conductor-conductor" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.857423 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e7a7c0-e906-4e2d-95e1-0062edf3ff36" containerName="nova-cell0-conductor-conductor" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.857452 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d0f44d4-a284-4c21-bea7-b964477081c4" containerName="nova-cell1-conductor-conductor" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.858115 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.875628 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.889486 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.920069 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.933623 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.943431 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.944772 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.946339 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 08:54:39 crc kubenswrapper[4935]: I1005 08:54:39.958319 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.023990 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.048701 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.049084 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbm6n\" (UniqueName: \"kubernetes.io/projected/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-kube-api-access-vbm6n\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.049166 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw8c4\" (UniqueName: \"kubernetes.io/projected/006b0ab2-2454-4507-92fd-9c314088c339-kube-api-access-cw8c4\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.049192 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.049230 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.049283 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.151363 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.151597 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.151634 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbm6n\" (UniqueName: \"kubernetes.io/projected/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-kube-api-access-vbm6n\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.151771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw8c4\" (UniqueName: \"kubernetes.io/projected/006b0ab2-2454-4507-92fd-9c314088c339-kube-api-access-cw8c4\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.151833 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.151932 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.157968 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.158680 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.159538 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.175734 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.178471 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw8c4\" (UniqueName: \"kubernetes.io/projected/006b0ab2-2454-4507-92fd-9c314088c339-kube-api-access-cw8c4\") pod \"nova-cell0-conductor-0\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.179397 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbm6n\" (UniqueName: \"kubernetes.io/projected/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-kube-api-access-vbm6n\") pod \"nova-cell1-conductor-0\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.219113 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.263403 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.342976 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:40 crc kubenswrapper[4935]: W1005 08:54:40.720406 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod006b0ab2_2454_4507_92fd_9c314088c339.slice/crio-9bb6e092776f8d3cb38e9015c6cd0f5cb812b531d1609c7bd027e46a361b0f26 WatchSource:0}: Error finding container 9bb6e092776f8d3cb38e9015c6cd0f5cb812b531d1609c7bd027e46a361b0f26: Status 404 returned error can't find the container with id 9bb6e092776f8d3cb38e9015c6cd0f5cb812b531d1609c7bd027e46a361b0f26 Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.724218 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.795470 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d0f44d4-a284-4c21-bea7-b964477081c4" path="/var/lib/kubelet/pods/2d0f44d4-a284-4c21-bea7-b964477081c4/volumes" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.798320 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76e7a7c0-e906-4e2d-95e1-0062edf3ff36" path="/var/lib/kubelet/pods/76e7a7c0-e906-4e2d-95e1-0062edf3ff36/volumes" Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.800474 4935 generic.go:334] "Generic (PLEG): container finished" podID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerID="354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2" exitCode=0 Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.800607 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5wdx" event={"ID":"a8775ed9-5817-4c33-87e6-a1119d4ccc7b","Type":"ContainerDied","Data":"354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2"} Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.808787 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"006b0ab2-2454-4507-92fd-9c314088c339","Type":"ContainerStarted","Data":"9bb6e092776f8d3cb38e9015c6cd0f5cb812b531d1609c7bd027e46a361b0f26"} Oct 05 08:54:40 crc kubenswrapper[4935]: W1005 08:54:40.843877 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7899e5aa_56b1_4016_8e50_97f8ddcfa18f.slice/crio-d705521a01f1301441db7718da23644c93d16a31cd3130cf6aa6b1910da8fbaf WatchSource:0}: Error finding container d705521a01f1301441db7718da23644c93d16a31cd3130cf6aa6b1910da8fbaf: Status 404 returned error can't find the container with id d705521a01f1301441db7718da23644c93d16a31cd3130cf6aa6b1910da8fbaf Oct 05 08:54:40 crc kubenswrapper[4935]: I1005 08:54:40.854313 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.821956 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7899e5aa-56b1-4016-8e50-97f8ddcfa18f","Type":"ContainerStarted","Data":"d91182927717e44b7a0b70a56d7b9b1546ac136032d8996d1c45922d842f753e"} Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.822539 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7899e5aa-56b1-4016-8e50-97f8ddcfa18f","Type":"ContainerStarted","Data":"d705521a01f1301441db7718da23644c93d16a31cd3130cf6aa6b1910da8fbaf"} Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.823557 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.827332 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5wdx" event={"ID":"a8775ed9-5817-4c33-87e6-a1119d4ccc7b","Type":"ContainerStarted","Data":"045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b"} Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.829640 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"006b0ab2-2454-4507-92fd-9c314088c339","Type":"ContainerStarted","Data":"aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06"} Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.830230 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.845191 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.845171879 podStartE2EDuration="2.845171879s" podCreationTimestamp="2025-10-05 08:54:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:54:41.839491679 +0000 UTC m=+7315.722118149" watchObservedRunningTime="2025-10-05 08:54:41.845171879 +0000 UTC m=+7315.727798349" Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.872777 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v5wdx" podStartSLOduration=2.322286191 podStartE2EDuration="4.872758741s" podCreationTimestamp="2025-10-05 08:54:37 +0000 UTC" firstStartedPulling="2025-10-05 08:54:38.715141074 +0000 UTC m=+7312.597767534" lastFinishedPulling="2025-10-05 08:54:41.265613614 +0000 UTC m=+7315.148240084" observedRunningTime="2025-10-05 08:54:41.865854948 +0000 UTC m=+7315.748481418" watchObservedRunningTime="2025-10-05 08:54:41.872758741 +0000 UTC m=+7315.755385211" Oct 05 08:54:41 crc kubenswrapper[4935]: I1005 08:54:41.889779 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.8897607020000002 podStartE2EDuration="2.889760702s" podCreationTimestamp="2025-10-05 08:54:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:54:41.881651477 +0000 UTC m=+7315.764277957" watchObservedRunningTime="2025-10-05 08:54:41.889760702 +0000 UTC m=+7315.772387172" Oct 05 08:54:42 crc kubenswrapper[4935]: I1005 08:54:42.322126 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:54:42 crc kubenswrapper[4935]: I1005 08:54:42.323331 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:54:45 crc kubenswrapper[4935]: I1005 08:54:45.024531 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 08:54:45 crc kubenswrapper[4935]: I1005 08:54:45.070322 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 08:54:45 crc kubenswrapper[4935]: I1005 08:54:45.251079 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 05 08:54:45 crc kubenswrapper[4935]: I1005 08:54:45.343145 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:45 crc kubenswrapper[4935]: I1005 08:54:45.354363 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:45 crc kubenswrapper[4935]: I1005 08:54:45.888332 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:54:45 crc kubenswrapper[4935]: I1005 08:54:45.926554 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 08:54:47 crc kubenswrapper[4935]: I1005 08:54:47.074247 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ts4mj"] Oct 05 08:54:47 crc kubenswrapper[4935]: I1005 08:54:47.083118 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ts4mj"] Oct 05 08:54:47 crc kubenswrapper[4935]: I1005 08:54:47.322264 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:54:47 crc kubenswrapper[4935]: I1005 08:54:47.322363 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:54:47 crc kubenswrapper[4935]: I1005 08:54:47.777698 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:47 crc kubenswrapper[4935]: I1005 08:54:47.777760 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:47 crc kubenswrapper[4935]: I1005 08:54:47.866273 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:47 crc kubenswrapper[4935]: I1005 08:54:47.966447 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:48 crc kubenswrapper[4935]: I1005 08:54:48.113141 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5wdx"] Oct 05 08:54:48 crc kubenswrapper[4935]: I1005 08:54:48.185539 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:54:48 crc kubenswrapper[4935]: I1005 08:54:48.185587 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:54:48 crc kubenswrapper[4935]: I1005 08:54:48.406120 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.95:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:54:48 crc kubenswrapper[4935]: I1005 08:54:48.406493 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.95:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:54:48 crc kubenswrapper[4935]: I1005 08:54:48.796421 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f38b74-c71b-4a9c-8ac7-4bb13037bc07" path="/var/lib/kubelet/pods/95f38b74-c71b-4a9c-8ac7-4bb13037bc07/volumes" Oct 05 08:54:49 crc kubenswrapper[4935]: I1005 08:54:49.275073 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.97:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:54:49 crc kubenswrapper[4935]: I1005 08:54:49.275152 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.97:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:54:49 crc kubenswrapper[4935]: I1005 08:54:49.922545 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v5wdx" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerName="registry-server" containerID="cri-o://045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b" gracePeriod=2 Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.298489 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.395715 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.477367 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-utilities\") pod \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.477421 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w4p6\" (UniqueName: \"kubernetes.io/projected/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-kube-api-access-7w4p6\") pod \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.478233 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-utilities" (OuterVolumeSpecName: "utilities") pod "a8775ed9-5817-4c33-87e6-a1119d4ccc7b" (UID: "a8775ed9-5817-4c33-87e6-a1119d4ccc7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.484351 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-kube-api-access-7w4p6" (OuterVolumeSpecName: "kube-api-access-7w4p6") pod "a8775ed9-5817-4c33-87e6-a1119d4ccc7b" (UID: "a8775ed9-5817-4c33-87e6-a1119d4ccc7b"). InnerVolumeSpecName "kube-api-access-7w4p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.578437 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-catalog-content\") pod \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\" (UID: \"a8775ed9-5817-4c33-87e6-a1119d4ccc7b\") " Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.592355 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.592402 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w4p6\" (UniqueName: \"kubernetes.io/projected/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-kube-api-access-7w4p6\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.645291 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8775ed9-5817-4c33-87e6-a1119d4ccc7b" (UID: "a8775ed9-5817-4c33-87e6-a1119d4ccc7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.694395 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8775ed9-5817-4c33-87e6-a1119d4ccc7b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.935986 4935 generic.go:334] "Generic (PLEG): container finished" podID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerID="045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b" exitCode=0 Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.936064 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5wdx" event={"ID":"a8775ed9-5817-4c33-87e6-a1119d4ccc7b","Type":"ContainerDied","Data":"045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b"} Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.936090 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5wdx" event={"ID":"a8775ed9-5817-4c33-87e6-a1119d4ccc7b","Type":"ContainerDied","Data":"4e5b4d196dbefc7997378a45c7e9545e8f17886ebbf418464d903e2973cc06dd"} Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.936108 4935 scope.go:117] "RemoveContainer" containerID="045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.936227 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5wdx" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.961155 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5wdx"] Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.975006 4935 scope.go:117] "RemoveContainer" containerID="354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2" Oct 05 08:54:50 crc kubenswrapper[4935]: I1005 08:54:50.977604 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v5wdx"] Oct 05 08:54:51 crc kubenswrapper[4935]: I1005 08:54:51.002077 4935 scope.go:117] "RemoveContainer" containerID="dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f" Oct 05 08:54:51 crc kubenswrapper[4935]: I1005 08:54:51.058304 4935 scope.go:117] "RemoveContainer" containerID="045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b" Oct 05 08:54:51 crc kubenswrapper[4935]: E1005 08:54:51.058823 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b\": container with ID starting with 045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b not found: ID does not exist" containerID="045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b" Oct 05 08:54:51 crc kubenswrapper[4935]: I1005 08:54:51.058880 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b"} err="failed to get container status \"045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b\": rpc error: code = NotFound desc = could not find container \"045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b\": container with ID starting with 045516b2b65efa610750c45747bfba0ebcac1e834303bb6aed0eab6b4f5b595b not found: ID does not exist" Oct 05 08:54:51 crc kubenswrapper[4935]: I1005 08:54:51.058934 4935 scope.go:117] "RemoveContainer" containerID="354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2" Oct 05 08:54:51 crc kubenswrapper[4935]: E1005 08:54:51.059472 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2\": container with ID starting with 354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2 not found: ID does not exist" containerID="354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2" Oct 05 08:54:51 crc kubenswrapper[4935]: I1005 08:54:51.059514 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2"} err="failed to get container status \"354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2\": rpc error: code = NotFound desc = could not find container \"354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2\": container with ID starting with 354ad6583f8c8c966cc364aec76025379b8bdcef235479dacc42b07c9d6f46e2 not found: ID does not exist" Oct 05 08:54:51 crc kubenswrapper[4935]: I1005 08:54:51.059542 4935 scope.go:117] "RemoveContainer" containerID="dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f" Oct 05 08:54:51 crc kubenswrapper[4935]: E1005 08:54:51.059835 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f\": container with ID starting with dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f not found: ID does not exist" containerID="dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f" Oct 05 08:54:51 crc kubenswrapper[4935]: I1005 08:54:51.059863 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f"} err="failed to get container status \"dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f\": rpc error: code = NotFound desc = could not find container \"dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f\": container with ID starting with dbe2530f49e9fed0bc3cf6aaffbeb7a08b6009f681cd608e916d9446f4f1c53f not found: ID does not exist" Oct 05 08:54:52 crc kubenswrapper[4935]: I1005 08:54:52.793794 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" path="/var/lib/kubelet/pods/a8775ed9-5817-4c33-87e6-a1119d4ccc7b/volumes" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.036196 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2441-account-create-c9nbz"] Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.055860 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2441-account-create-c9nbz"] Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.345746 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.383674 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.388329 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.626567 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:54:57 crc kubenswrapper[4935]: E1005 08:54:57.627053 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerName="extract-utilities" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.627075 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerName="extract-utilities" Oct 05 08:54:57 crc kubenswrapper[4935]: E1005 08:54:57.627097 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerName="registry-server" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.627106 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerName="registry-server" Oct 05 08:54:57 crc kubenswrapper[4935]: E1005 08:54:57.627135 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerName="extract-content" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.627144 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerName="extract-content" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.627392 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8775ed9-5817-4c33-87e6-a1119d4ccc7b" containerName="registry-server" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.628636 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.631267 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.671361 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.735265 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-scripts\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.735514 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.735919 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw6j6\" (UniqueName: \"kubernetes.io/projected/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-kube-api-access-sw6j6\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.735979 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.736182 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.736230 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.838010 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw6j6\" (UniqueName: \"kubernetes.io/projected/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-kube-api-access-sw6j6\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.838093 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.838143 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.838217 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.838317 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-scripts\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.838409 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.838636 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.846356 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.847382 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.847622 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-scripts\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.856291 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.858423 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw6j6\" (UniqueName: \"kubernetes.io/projected/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-kube-api-access-sw6j6\") pod \"cinder-scheduler-0\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " pod="openstack/cinder-scheduler-0" Oct 05 08:54:57 crc kubenswrapper[4935]: I1005 08:54:57.946480 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:54:58 crc kubenswrapper[4935]: I1005 08:54:58.034136 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 08:54:58 crc kubenswrapper[4935]: I1005 08:54:58.190797 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 08:54:58 crc kubenswrapper[4935]: I1005 08:54:58.191739 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:54:58 crc kubenswrapper[4935]: I1005 08:54:58.191995 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 08:54:58 crc kubenswrapper[4935]: I1005 08:54:58.195103 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 08:54:58 crc kubenswrapper[4935]: I1005 08:54:58.440370 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:54:58 crc kubenswrapper[4935]: I1005 08:54:58.790913 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23135322-e76e-457d-82d0-10463fefd374" path="/var/lib/kubelet/pods/23135322-e76e-457d-82d0-10463fefd374/volumes" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.043599 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc","Type":"ContainerStarted","Data":"b201e3752d763e0a57e7cde0d81e073987207497e4303147390cdd1e4f24872b"} Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.044005 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.060830 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.182019 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.182319 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api-log" containerID="cri-o://1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb" gracePeriod=30 Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.182472 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api" containerID="cri-o://a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f" gracePeriod=30 Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.680541 4935 scope.go:117] "RemoveContainer" containerID="21f5649b65aa66bc0cdb34ebf3628b79e4a6d6fd33b08888e7cbab3ce5b67c40" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.721958 4935 scope.go:117] "RemoveContainer" containerID="3e58520a17cf1fe6c800531870e7f62071ac707bde020a522d6323c2b57f29ee" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.760624 4935 scope.go:117] "RemoveContainer" containerID="759caccedc95fc23d130fab32680183f1cea064f59ff7f38b6b83195c0cd4a22" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.783866 4935 scope.go:117] "RemoveContainer" containerID="c7055ef8b0d480c8b737e73faeb9be4df7153079b1e0fcb0fe9d395f44d108d1" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.944825 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.946935 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.950077 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 05 08:54:59 crc kubenswrapper[4935]: I1005 08:54:59.956066 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.076151 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc","Type":"ContainerStarted","Data":"2a2fdada3c63c319a4f6a86bb3c0edec1a58f749218e3d85e7ab4d1776e80965"} Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.079467 4935 generic.go:334] "Generic (PLEG): container finished" podID="8fd19384-485b-4224-ae32-5f54319214d4" containerID="1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb" exitCode=143 Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.080247 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8fd19384-485b-4224-ae32-5f54319214d4","Type":"ContainerDied","Data":"1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb"} Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.082960 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.082992 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083057 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083080 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083096 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-run\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083131 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptm8f\" (UniqueName: \"kubernetes.io/projected/490ba015-eba1-4f8e-a935-f8d27de1dc8d-kube-api-access-ptm8f\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083162 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083186 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083220 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083238 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083261 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083284 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/490ba015-eba1-4f8e-a935-f8d27de1dc8d-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083298 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083313 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083331 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.083358 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184715 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptm8f\" (UniqueName: \"kubernetes.io/projected/490ba015-eba1-4f8e-a935-f8d27de1dc8d-kube-api-access-ptm8f\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184786 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184807 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184852 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184869 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184901 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184923 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/490ba015-eba1-4f8e-a935-f8d27de1dc8d-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184948 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184964 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.184981 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.185027 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.185077 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.185095 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.185121 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.185143 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.185158 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-run\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.185165 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.185235 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-run\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.186345 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.187495 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.187526 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.187607 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.186874 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.188370 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.188425 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.188535 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/490ba015-eba1-4f8e-a935-f8d27de1dc8d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.190969 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/490ba015-eba1-4f8e-a935-f8d27de1dc8d-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.191048 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.191432 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.191958 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.199544 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/490ba015-eba1-4f8e-a935-f8d27de1dc8d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.204668 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptm8f\" (UniqueName: \"kubernetes.io/projected/490ba015-eba1-4f8e-a935-f8d27de1dc8d-kube-api-access-ptm8f\") pod \"cinder-volume-volume1-0\" (UID: \"490ba015-eba1-4f8e-a935-f8d27de1dc8d\") " pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.277062 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.469675 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.472331 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.476629 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.478688 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591178 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-sys\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591244 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591262 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591313 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591333 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cb2a652-0e57-4699-85d5-69e6966e8948-ceph\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591352 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg4xm\" (UniqueName: \"kubernetes.io/projected/7cb2a652-0e57-4699-85d5-69e6966e8948-kube-api-access-mg4xm\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591433 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-scripts\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591474 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591489 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-run\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591515 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591562 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591584 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591604 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-lib-modules\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591634 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-dev\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591652 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-config-data\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.591680 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.693273 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.693335 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-dev\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.693352 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-lib-modules\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.693370 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-config-data\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.693404 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.693436 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-dev\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694070 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-sys\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694110 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-sys\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694127 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-lib-modules\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694161 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694196 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694215 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694249 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694286 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cb2a652-0e57-4699-85d5-69e6966e8948-ceph\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694303 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg4xm\" (UniqueName: \"kubernetes.io/projected/7cb2a652-0e57-4699-85d5-69e6966e8948-kube-api-access-mg4xm\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694372 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-scripts\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694395 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694410 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-run\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694436 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.694464 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.695182 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.695256 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.695290 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.696438 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.696458 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-run\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.696504 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7cb2a652-0e57-4699-85d5-69e6966e8948-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.697867 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.698519 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-config-data\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.701749 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-scripts\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.702180 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cb2a652-0e57-4699-85d5-69e6966e8948-ceph\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.702239 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cb2a652-0e57-4699-85d5-69e6966e8948-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.711507 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg4xm\" (UniqueName: \"kubernetes.io/projected/7cb2a652-0e57-4699-85d5-69e6966e8948-kube-api-access-mg4xm\") pod \"cinder-backup-0\" (UID: \"7cb2a652-0e57-4699-85d5-69e6966e8948\") " pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.805154 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 05 08:55:00 crc kubenswrapper[4935]: I1005 08:55:00.864982 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 05 08:55:00 crc kubenswrapper[4935]: W1005 08:55:00.865383 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod490ba015_eba1_4f8e_a935_f8d27de1dc8d.slice/crio-50c1299782fd042dd3b394d7fddcf41758a4fd3c41f1a3d2d4afb14cce51e592 WatchSource:0}: Error finding container 50c1299782fd042dd3b394d7fddcf41758a4fd3c41f1a3d2d4afb14cce51e592: Status 404 returned error can't find the container with id 50c1299782fd042dd3b394d7fddcf41758a4fd3c41f1a3d2d4afb14cce51e592 Oct 05 08:55:01 crc kubenswrapper[4935]: I1005 08:55:01.098358 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc","Type":"ContainerStarted","Data":"e3a7174cdb0db5bb80415071b1b75630bcef265e3fbb14e06e512bfe625a27c1"} Oct 05 08:55:01 crc kubenswrapper[4935]: I1005 08:55:01.101159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"490ba015-eba1-4f8e-a935-f8d27de1dc8d","Type":"ContainerStarted","Data":"50c1299782fd042dd3b394d7fddcf41758a4fd3c41f1a3d2d4afb14cce51e592"} Oct 05 08:55:01 crc kubenswrapper[4935]: I1005 08:55:01.117966 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.8057544229999998 podStartE2EDuration="4.117949361s" podCreationTimestamp="2025-10-05 08:54:57 +0000 UTC" firstStartedPulling="2025-10-05 08:54:58.457410663 +0000 UTC m=+7332.340037123" lastFinishedPulling="2025-10-05 08:54:58.769605601 +0000 UTC m=+7332.652232061" observedRunningTime="2025-10-05 08:55:01.115241669 +0000 UTC m=+7334.997868149" watchObservedRunningTime="2025-10-05 08:55:01.117949361 +0000 UTC m=+7335.000575821" Oct 05 08:55:01 crc kubenswrapper[4935]: I1005 08:55:01.387418 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.121475 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"490ba015-eba1-4f8e-a935-f8d27de1dc8d","Type":"ContainerStarted","Data":"4e3d9f58065f2838fd0872029f75517983405ddda6a2f0f5505d45d1591fc779"} Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.122038 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"490ba015-eba1-4f8e-a935-f8d27de1dc8d","Type":"ContainerStarted","Data":"a8f1a376ff37c56b52365cdc6e5a780f51b20eb0be76ea760745a0b22d5fcc28"} Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.123578 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"7cb2a652-0e57-4699-85d5-69e6966e8948","Type":"ContainerStarted","Data":"bd08b7114d6180778b14d8c8b3312ab956fd553b48117515ab02e4e74cd448d7"} Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.123620 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"7cb2a652-0e57-4699-85d5-69e6966e8948","Type":"ContainerStarted","Data":"411d362ce580c78b1ce0cb456b92cffa4fa17875432a48458a0e72aca780ed67"} Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.158844 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.798099943 podStartE2EDuration="3.158809197s" podCreationTimestamp="2025-10-05 08:54:59 +0000 UTC" firstStartedPulling="2025-10-05 08:55:00.867671365 +0000 UTC m=+7334.750297825" lastFinishedPulling="2025-10-05 08:55:01.228380619 +0000 UTC m=+7335.111007079" observedRunningTime="2025-10-05 08:55:02.150181359 +0000 UTC m=+7336.032807839" watchObservedRunningTime="2025-10-05 08:55:02.158809197 +0000 UTC m=+7336.041435657" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.358678 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.92:8776/healthcheck\": read tcp 10.217.0.2:51306->10.217.1.92:8776: read: connection reset by peer" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.685435 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.843743 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-scripts\") pod \"8fd19384-485b-4224-ae32-5f54319214d4\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.843782 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fd19384-485b-4224-ae32-5f54319214d4-logs\") pod \"8fd19384-485b-4224-ae32-5f54319214d4\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.843809 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8fd19384-485b-4224-ae32-5f54319214d4-etc-machine-id\") pod \"8fd19384-485b-4224-ae32-5f54319214d4\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.843862 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv5s5\" (UniqueName: \"kubernetes.io/projected/8fd19384-485b-4224-ae32-5f54319214d4-kube-api-access-jv5s5\") pod \"8fd19384-485b-4224-ae32-5f54319214d4\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.843903 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-combined-ca-bundle\") pod \"8fd19384-485b-4224-ae32-5f54319214d4\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.843936 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data\") pod \"8fd19384-485b-4224-ae32-5f54319214d4\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.843970 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data-custom\") pod \"8fd19384-485b-4224-ae32-5f54319214d4\" (UID: \"8fd19384-485b-4224-ae32-5f54319214d4\") " Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.844370 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fd19384-485b-4224-ae32-5f54319214d4-logs" (OuterVolumeSpecName: "logs") pod "8fd19384-485b-4224-ae32-5f54319214d4" (UID: "8fd19384-485b-4224-ae32-5f54319214d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.844726 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fd19384-485b-4224-ae32-5f54319214d4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8fd19384-485b-4224-ae32-5f54319214d4" (UID: "8fd19384-485b-4224-ae32-5f54319214d4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.845419 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fd19384-485b-4224-ae32-5f54319214d4-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.845465 4935 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8fd19384-485b-4224-ae32-5f54319214d4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.849153 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-scripts" (OuterVolumeSpecName: "scripts") pod "8fd19384-485b-4224-ae32-5f54319214d4" (UID: "8fd19384-485b-4224-ae32-5f54319214d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.849338 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8fd19384-485b-4224-ae32-5f54319214d4" (UID: "8fd19384-485b-4224-ae32-5f54319214d4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.850551 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd19384-485b-4224-ae32-5f54319214d4-kube-api-access-jv5s5" (OuterVolumeSpecName: "kube-api-access-jv5s5") pod "8fd19384-485b-4224-ae32-5f54319214d4" (UID: "8fd19384-485b-4224-ae32-5f54319214d4"). InnerVolumeSpecName "kube-api-access-jv5s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.889695 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fd19384-485b-4224-ae32-5f54319214d4" (UID: "8fd19384-485b-4224-ae32-5f54319214d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.911983 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data" (OuterVolumeSpecName: "config-data") pod "8fd19384-485b-4224-ae32-5f54319214d4" (UID: "8fd19384-485b-4224-ae32-5f54319214d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.947119 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.947322 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.947357 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv5s5\" (UniqueName: \"kubernetes.io/projected/8fd19384-485b-4224-ae32-5f54319214d4-kube-api-access-jv5s5\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.947374 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.947384 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:02 crc kubenswrapper[4935]: I1005 08:55:02.947396 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd19384-485b-4224-ae32-5f54319214d4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.136548 4935 generic.go:334] "Generic (PLEG): container finished" podID="8fd19384-485b-4224-ae32-5f54319214d4" containerID="a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f" exitCode=0 Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.138486 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8fd19384-485b-4224-ae32-5f54319214d4","Type":"ContainerDied","Data":"a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f"} Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.138531 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8fd19384-485b-4224-ae32-5f54319214d4","Type":"ContainerDied","Data":"82c3ed077af29a72a8116b958bf438b7e2c84ab4c3d3c33aa01fc22a142cf4c9"} Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.138555 4935 scope.go:117] "RemoveContainer" containerID="a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.139037 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.150704 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"7cb2a652-0e57-4699-85d5-69e6966e8948","Type":"ContainerStarted","Data":"d9ee1919ebf4f6647ce79d3a71e84060d60ea3b6193988b68e66744291e5667e"} Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.177121 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.820298785 podStartE2EDuration="3.177100455s" podCreationTimestamp="2025-10-05 08:55:00 +0000 UTC" firstStartedPulling="2025-10-05 08:55:01.391711609 +0000 UTC m=+7335.274338069" lastFinishedPulling="2025-10-05 08:55:01.748513279 +0000 UTC m=+7335.631139739" observedRunningTime="2025-10-05 08:55:03.176373746 +0000 UTC m=+7337.059000226" watchObservedRunningTime="2025-10-05 08:55:03.177100455 +0000 UTC m=+7337.059726915" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.181965 4935 scope.go:117] "RemoveContainer" containerID="1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.202938 4935 scope.go:117] "RemoveContainer" containerID="a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f" Oct 05 08:55:03 crc kubenswrapper[4935]: E1005 08:55:03.203857 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f\": container with ID starting with a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f not found: ID does not exist" containerID="a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.203957 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f"} err="failed to get container status \"a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f\": rpc error: code = NotFound desc = could not find container \"a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f\": container with ID starting with a5cbaec23ad2026b05150825b27d878a17a379fce71d3205ea3be104aebc9b1f not found: ID does not exist" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.204020 4935 scope.go:117] "RemoveContainer" containerID="1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb" Oct 05 08:55:03 crc kubenswrapper[4935]: E1005 08:55:03.212432 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb\": container with ID starting with 1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb not found: ID does not exist" containerID="1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.212688 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb"} err="failed to get container status \"1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb\": rpc error: code = NotFound desc = could not find container \"1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb\": container with ID starting with 1177c25799228e0b282ec5d96ef4f64802dca80de7badd18510554af968a94bb not found: ID does not exist" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.222800 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.232729 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.250812 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:55:03 crc kubenswrapper[4935]: E1005 08:55:03.251476 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api-log" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.251499 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api-log" Oct 05 08:55:03 crc kubenswrapper[4935]: E1005 08:55:03.251511 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.251522 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.251771 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.251792 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd19384-485b-4224-ae32-5f54319214d4" containerName="cinder-api-log" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.253817 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.256543 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-config-data-custom\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.256587 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-scripts\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.256677 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdk7r\" (UniqueName: \"kubernetes.io/projected/9cec9943-9a6f-4752-b334-f49d622c1929-kube-api-access-tdk7r\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.256760 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cec9943-9a6f-4752-b334-f49d622c1929-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.256812 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-config-data\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.256832 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.256956 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cec9943-9a6f-4752-b334-f49d622c1929-logs\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.259424 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.286011 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.358748 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdk7r\" (UniqueName: \"kubernetes.io/projected/9cec9943-9a6f-4752-b334-f49d622c1929-kube-api-access-tdk7r\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.359181 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cec9943-9a6f-4752-b334-f49d622c1929-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.359288 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-config-data\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.359579 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cec9943-9a6f-4752-b334-f49d622c1929-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.360341 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.360423 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cec9943-9a6f-4752-b334-f49d622c1929-logs\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.360564 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-config-data-custom\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.360602 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-scripts\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.361140 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cec9943-9a6f-4752-b334-f49d622c1929-logs\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.363780 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.364355 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-config-data\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.365696 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-scripts\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.372969 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cec9943-9a6f-4752-b334-f49d622c1929-config-data-custom\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.378658 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdk7r\" (UniqueName: \"kubernetes.io/projected/9cec9943-9a6f-4752-b334-f49d622c1929-kube-api-access-tdk7r\") pod \"cinder-api-0\" (UID: \"9cec9943-9a6f-4752-b334-f49d622c1929\") " pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.592700 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:55:03 crc kubenswrapper[4935]: I1005 08:55:03.914546 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:55:03 crc kubenswrapper[4935]: W1005 08:55:03.923419 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cec9943_9a6f_4752_b334_f49d622c1929.slice/crio-ae80b7c1935e5b0ff3942dfd9e282c00e9eea51dacc4ff572b70bfb2ff8b717c WatchSource:0}: Error finding container ae80b7c1935e5b0ff3942dfd9e282c00e9eea51dacc4ff572b70bfb2ff8b717c: Status 404 returned error can't find the container with id ae80b7c1935e5b0ff3942dfd9e282c00e9eea51dacc4ff572b70bfb2ff8b717c Oct 05 08:55:04 crc kubenswrapper[4935]: I1005 08:55:04.168800 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9cec9943-9a6f-4752-b334-f49d622c1929","Type":"ContainerStarted","Data":"ae80b7c1935e5b0ff3942dfd9e282c00e9eea51dacc4ff572b70bfb2ff8b717c"} Oct 05 08:55:04 crc kubenswrapper[4935]: I1005 08:55:04.816098 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd19384-485b-4224-ae32-5f54319214d4" path="/var/lib/kubelet/pods/8fd19384-485b-4224-ae32-5f54319214d4/volumes" Oct 05 08:55:05 crc kubenswrapper[4935]: I1005 08:55:05.178100 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9cec9943-9a6f-4752-b334-f49d622c1929","Type":"ContainerStarted","Data":"3647e3977f2942fc7efad9f22bc588639aaeaafb356767c2e008c7d7d83ffe13"} Oct 05 08:55:05 crc kubenswrapper[4935]: I1005 08:55:05.179663 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 08:55:05 crc kubenswrapper[4935]: I1005 08:55:05.179704 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9cec9943-9a6f-4752-b334-f49d622c1929","Type":"ContainerStarted","Data":"cb2ffc403617cef564fe80a06db72e9b61aa0ec1dc8ad9dec4089dd8b975892f"} Oct 05 08:55:05 crc kubenswrapper[4935]: I1005 08:55:05.207823 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.207799454 podStartE2EDuration="2.207799454s" podCreationTimestamp="2025-10-05 08:55:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:55:05.197934742 +0000 UTC m=+7339.080561222" watchObservedRunningTime="2025-10-05 08:55:05.207799454 +0000 UTC m=+7339.090425934" Oct 05 08:55:05 crc kubenswrapper[4935]: I1005 08:55:05.277725 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:05 crc kubenswrapper[4935]: I1005 08:55:05.805693 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 05 08:55:08 crc kubenswrapper[4935]: I1005 08:55:08.199585 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 08:55:08 crc kubenswrapper[4935]: I1005 08:55:08.297359 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:55:09 crc kubenswrapper[4935]: I1005 08:55:09.059922 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-g7kbv"] Oct 05 08:55:09 crc kubenswrapper[4935]: I1005 08:55:09.074235 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-g7kbv"] Oct 05 08:55:09 crc kubenswrapper[4935]: I1005 08:55:09.217576 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerName="probe" containerID="cri-o://e3a7174cdb0db5bb80415071b1b75630bcef265e3fbb14e06e512bfe625a27c1" gracePeriod=30 Oct 05 08:55:09 crc kubenswrapper[4935]: I1005 08:55:09.217810 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerName="cinder-scheduler" containerID="cri-o://2a2fdada3c63c319a4f6a86bb3c0edec1a58f749218e3d85e7ab4d1776e80965" gracePeriod=30 Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.233693 4935 generic.go:334] "Generic (PLEG): container finished" podID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerID="e3a7174cdb0db5bb80415071b1b75630bcef265e3fbb14e06e512bfe625a27c1" exitCode=0 Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.234108 4935 generic.go:334] "Generic (PLEG): container finished" podID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerID="2a2fdada3c63c319a4f6a86bb3c0edec1a58f749218e3d85e7ab4d1776e80965" exitCode=0 Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.234129 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc","Type":"ContainerDied","Data":"e3a7174cdb0db5bb80415071b1b75630bcef265e3fbb14e06e512bfe625a27c1"} Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.234155 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc","Type":"ContainerDied","Data":"2a2fdada3c63c319a4f6a86bb3c0edec1a58f749218e3d85e7ab4d1776e80965"} Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.497676 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.602928 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.725869 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw6j6\" (UniqueName: \"kubernetes.io/projected/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-kube-api-access-sw6j6\") pod \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.726367 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-combined-ca-bundle\") pod \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.726478 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data-custom\") pod \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.726524 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-scripts\") pod \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.726784 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-etc-machine-id\") pod \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.726864 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data\") pod \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\" (UID: \"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc\") " Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.728468 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" (UID: "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.733286 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" (UID: "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.733805 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-kube-api-access-sw6j6" (OuterVolumeSpecName: "kube-api-access-sw6j6") pod "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" (UID: "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc"). InnerVolumeSpecName "kube-api-access-sw6j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.736008 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-scripts" (OuterVolumeSpecName: "scripts") pod "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" (UID: "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.791214 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308d5101-f413-4fd3-b5fa-90450a246f5f" path="/var/lib/kubelet/pods/308d5101-f413-4fd3-b5fa-90450a246f5f/volumes" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.804784 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" (UID: "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.830034 4935 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.830079 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw6j6\" (UniqueName: \"kubernetes.io/projected/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-kube-api-access-sw6j6\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.830091 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.830101 4935 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.830111 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.853620 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data" (OuterVolumeSpecName: "config-data") pod "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" (UID: "472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:55:10 crc kubenswrapper[4935]: I1005 08:55:10.932625 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.090281 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.250755 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc","Type":"ContainerDied","Data":"b201e3752d763e0a57e7cde0d81e073987207497e4303147390cdd1e4f24872b"} Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.250828 4935 scope.go:117] "RemoveContainer" containerID="e3a7174cdb0db5bb80415071b1b75630bcef265e3fbb14e06e512bfe625a27c1" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.250878 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.298175 4935 scope.go:117] "RemoveContainer" containerID="2a2fdada3c63c319a4f6a86bb3c0edec1a58f749218e3d85e7ab4d1776e80965" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.309653 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.334026 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.346047 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:55:11 crc kubenswrapper[4935]: E1005 08:55:11.346645 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerName="probe" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.346677 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerName="probe" Oct 05 08:55:11 crc kubenswrapper[4935]: E1005 08:55:11.346694 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerName="cinder-scheduler" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.346706 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerName="cinder-scheduler" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.346988 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerName="probe" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.347016 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" containerName="cinder-scheduler" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.348320 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.352359 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.352861 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.440210 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cx25\" (UniqueName: \"kubernetes.io/projected/2ae0b7bf-20f9-4942-898f-d44b8c061d30-kube-api-access-4cx25\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.440263 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-scripts\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.440320 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-config-data\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.440357 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.440708 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ae0b7bf-20f9-4942-898f-d44b8c061d30-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.440764 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.542933 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cx25\" (UniqueName: \"kubernetes.io/projected/2ae0b7bf-20f9-4942-898f-d44b8c061d30-kube-api-access-4cx25\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.542993 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-scripts\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.543020 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-config-data\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.543055 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.543211 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ae0b7bf-20f9-4942-898f-d44b8c061d30-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.543240 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.543342 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ae0b7bf-20f9-4942-898f-d44b8c061d30-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.548493 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.549455 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-scripts\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.549594 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.553080 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae0b7bf-20f9-4942-898f-d44b8c061d30-config-data\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.564070 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cx25\" (UniqueName: \"kubernetes.io/projected/2ae0b7bf-20f9-4942-898f-d44b8c061d30-kube-api-access-4cx25\") pod \"cinder-scheduler-0\" (UID: \"2ae0b7bf-20f9-4942-898f-d44b8c061d30\") " pod="openstack/cinder-scheduler-0" Oct 05 08:55:11 crc kubenswrapper[4935]: I1005 08:55:11.671377 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:55:12 crc kubenswrapper[4935]: I1005 08:55:12.224715 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:55:12 crc kubenswrapper[4935]: I1005 08:55:12.263642 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2ae0b7bf-20f9-4942-898f-d44b8c061d30","Type":"ContainerStarted","Data":"a730b61c04d575829ca1affd00651e8312f51842edc7e85b3730554ed1d97f39"} Oct 05 08:55:12 crc kubenswrapper[4935]: I1005 08:55:12.792018 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc" path="/var/lib/kubelet/pods/472f0a7f-8fdc-416b-9dfc-4a4c24d0b1bc/volumes" Oct 05 08:55:13 crc kubenswrapper[4935]: I1005 08:55:13.281522 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2ae0b7bf-20f9-4942-898f-d44b8c061d30","Type":"ContainerStarted","Data":"48abea7b4929d62c89e45061cab6babedc2a86da29290983a5f420f39bf54ebe"} Oct 05 08:55:14 crc kubenswrapper[4935]: I1005 08:55:14.301162 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2ae0b7bf-20f9-4942-898f-d44b8c061d30","Type":"ContainerStarted","Data":"949903d41a8580c495f93448087f7c82c35edffa74f953a7a992f5268b963584"} Oct 05 08:55:14 crc kubenswrapper[4935]: I1005 08:55:14.328613 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.328594381 podStartE2EDuration="3.328594381s" podCreationTimestamp="2025-10-05 08:55:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:55:14.328356335 +0000 UTC m=+7348.210982835" watchObservedRunningTime="2025-10-05 08:55:14.328594381 +0000 UTC m=+7348.211220851" Oct 05 08:55:15 crc kubenswrapper[4935]: I1005 08:55:15.490985 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 08:55:16 crc kubenswrapper[4935]: I1005 08:55:16.671908 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 08:55:22 crc kubenswrapper[4935]: I1005 08:55:22.053155 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-t7nvk"] Oct 05 08:55:22 crc kubenswrapper[4935]: I1005 08:55:22.059979 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-t7nvk"] Oct 05 08:55:22 crc kubenswrapper[4935]: I1005 08:55:22.122616 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 08:55:22 crc kubenswrapper[4935]: I1005 08:55:22.788372 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b4f125-73af-4d84-bc8e-a11257c7df45" path="/var/lib/kubelet/pods/37b4f125-73af-4d84-bc8e-a11257c7df45/volumes" Oct 05 08:56:00 crc kubenswrapper[4935]: I1005 08:56:00.079140 4935 scope.go:117] "RemoveContainer" containerID="1c97190c183f6cd258b44028c533845f0d0cba070667fbd128b80f06aa6301bf" Oct 05 08:56:00 crc kubenswrapper[4935]: I1005 08:56:00.133308 4935 scope.go:117] "RemoveContainer" containerID="8c00a3535c0c2b6e91f5b5aa2f990070f33798c6675ceb34c83653f2ce734e14" Oct 05 08:56:44 crc kubenswrapper[4935]: I1005 08:56:44.290441 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:56:44 crc kubenswrapper[4935]: I1005 08:56:44.291233 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.605606 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bc74c48f7-mfk2x"] Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.608383 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.613580 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.613644 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.613837 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-z29dx" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.614285 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.625877 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bc74c48f7-mfk2x"] Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.673498 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.673732 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerName="glance-log" containerID="cri-o://9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb" gracePeriod=30 Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.673866 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerName="glance-httpd" containerID="cri-o://c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385" gracePeriod=30 Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.717270 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-config-data\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.717324 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4zwd\" (UniqueName: \"kubernetes.io/projected/da554f1c-7227-4af1-991d-e9431ba90e68-kube-api-access-c4zwd\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.717385 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da554f1c-7227-4af1-991d-e9431ba90e68-horizon-secret-key\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.717415 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-scripts\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.717435 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da554f1c-7227-4af1-991d-e9431ba90e68-logs\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.736751 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66575b9d69-5f2nd"] Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.744793 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.749553 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66575b9d69-5f2nd"] Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.774275 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.774517 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerName="glance-httpd" containerID="cri-o://7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b" gracePeriod=30 Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.774557 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerName="glance-log" containerID="cri-o://62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613" gracePeriod=30 Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.819427 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-scripts\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.819806 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-config-data\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.820011 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-config-data\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.820064 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4zwd\" (UniqueName: \"kubernetes.io/projected/da554f1c-7227-4af1-991d-e9431ba90e68-kube-api-access-c4zwd\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.820465 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b87908c-7e5c-402d-80bf-954d6a601f64-horizon-secret-key\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.820546 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b87908c-7e5c-402d-80bf-954d6a601f64-logs\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.820598 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/1b87908c-7e5c-402d-80bf-954d6a601f64-kube-api-access-pfrsf\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.820637 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da554f1c-7227-4af1-991d-e9431ba90e68-horizon-secret-key\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.820743 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-scripts\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.820769 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da554f1c-7227-4af1-991d-e9431ba90e68-logs\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.821320 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-scripts\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.821690 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-config-data\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.821987 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da554f1c-7227-4af1-991d-e9431ba90e68-logs\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.835380 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da554f1c-7227-4af1-991d-e9431ba90e68-horizon-secret-key\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.837362 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4zwd\" (UniqueName: \"kubernetes.io/projected/da554f1c-7227-4af1-991d-e9431ba90e68-kube-api-access-c4zwd\") pod \"horizon-7bc74c48f7-mfk2x\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.922511 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b87908c-7e5c-402d-80bf-954d6a601f64-horizon-secret-key\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.922625 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b87908c-7e5c-402d-80bf-954d6a601f64-logs\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.922654 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/1b87908c-7e5c-402d-80bf-954d6a601f64-kube-api-access-pfrsf\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.922737 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-scripts\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.922756 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-config-data\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.923561 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b87908c-7e5c-402d-80bf-954d6a601f64-logs\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.923680 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-scripts\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.924311 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-config-data\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.927757 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b87908c-7e5c-402d-80bf-954d6a601f64-horizon-secret-key\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.935013 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:56:55 crc kubenswrapper[4935]: I1005 08:56:55.938594 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/1b87908c-7e5c-402d-80bf-954d6a601f64-kube-api-access-pfrsf\") pod \"horizon-66575b9d69-5f2nd\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.061320 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.216024 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66575b9d69-5f2nd"] Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.272957 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-75bb459497-g9b7r"] Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.274764 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.310094 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-75bb459497-g9b7r"] Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.360166 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm4np\" (UniqueName: \"kubernetes.io/projected/2cdeadef-85fc-4477-8352-dc671812251e-kube-api-access-lm4np\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.360454 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cdeadef-85fc-4477-8352-dc671812251e-logs\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.360625 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-config-data\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.360827 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-scripts\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.360939 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cdeadef-85fc-4477-8352-dc671812251e-horizon-secret-key\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.463056 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm4np\" (UniqueName: \"kubernetes.io/projected/2cdeadef-85fc-4477-8352-dc671812251e-kube-api-access-lm4np\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.463220 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cdeadef-85fc-4477-8352-dc671812251e-logs\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.463261 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-config-data\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.463316 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-scripts\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.463341 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cdeadef-85fc-4477-8352-dc671812251e-horizon-secret-key\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.463678 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cdeadef-85fc-4477-8352-dc671812251e-logs\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.464555 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-scripts\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.465848 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-config-data\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.471389 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cdeadef-85fc-4477-8352-dc671812251e-horizon-secret-key\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.490332 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm4np\" (UniqueName: \"kubernetes.io/projected/2cdeadef-85fc-4477-8352-dc671812251e-kube-api-access-lm4np\") pod \"horizon-75bb459497-g9b7r\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.556833 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bc74c48f7-mfk2x"] Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.557289 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"421b8438-b398-4a68-a88b-cf875d82bf1e","Type":"ContainerDied","Data":"62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613"} Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.557088 4935 generic.go:334] "Generic (PLEG): container finished" podID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerID="62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613" exitCode=143 Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.558961 4935 generic.go:334] "Generic (PLEG): container finished" podID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerID="9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb" exitCode=143 Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.559001 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9","Type":"ContainerDied","Data":"9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb"} Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.712347 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:56:56 crc kubenswrapper[4935]: I1005 08:56:56.848148 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66575b9d69-5f2nd"] Oct 05 08:56:57 crc kubenswrapper[4935]: I1005 08:56:57.179701 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-75bb459497-g9b7r"] Oct 05 08:56:57 crc kubenswrapper[4935]: W1005 08:56:57.190562 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cdeadef_85fc_4477_8352_dc671812251e.slice/crio-4a55228b78f7604e85790ff6894ab2d90ad03b932d74972fecab174d8200c6d1 WatchSource:0}: Error finding container 4a55228b78f7604e85790ff6894ab2d90ad03b932d74972fecab174d8200c6d1: Status 404 returned error can't find the container with id 4a55228b78f7604e85790ff6894ab2d90ad03b932d74972fecab174d8200c6d1 Oct 05 08:56:57 crc kubenswrapper[4935]: I1005 08:56:57.574932 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bb459497-g9b7r" event={"ID":"2cdeadef-85fc-4477-8352-dc671812251e","Type":"ContainerStarted","Data":"4a55228b78f7604e85790ff6894ab2d90ad03b932d74972fecab174d8200c6d1"} Oct 05 08:56:57 crc kubenswrapper[4935]: I1005 08:56:57.577322 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc74c48f7-mfk2x" event={"ID":"da554f1c-7227-4af1-991d-e9431ba90e68","Type":"ContainerStarted","Data":"22e87679d902ea723abac1c6b63278c1987feb3c6923a1a8b79ed91dbecd3c16"} Oct 05 08:56:57 crc kubenswrapper[4935]: I1005 08:56:57.580990 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66575b9d69-5f2nd" event={"ID":"1b87908c-7e5c-402d-80bf-954d6a601f64","Type":"ContainerStarted","Data":"d820bdc88b26636bbc22a8bf227792c4ccb1eade1f733577e35a729c0cacb017"} Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.434819 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.440731 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526441 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-ceph\") pod \"421b8438-b398-4a68-a88b-cf875d82bf1e\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526503 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-logs\") pod \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526551 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-config-data\") pod \"421b8438-b398-4a68-a88b-cf875d82bf1e\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526574 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-httpd-run\") pod \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526596 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-httpd-run\") pod \"421b8438-b398-4a68-a88b-cf875d82bf1e\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526619 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmdf4\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-kube-api-access-hmdf4\") pod \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526651 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-scripts\") pod \"421b8438-b398-4a68-a88b-cf875d82bf1e\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526722 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-config-data\") pod \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526740 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-scripts\") pod \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526778 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-combined-ca-bundle\") pod \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526802 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-ceph\") pod \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\" (UID: \"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526851 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf69g\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-kube-api-access-kf69g\") pod \"421b8438-b398-4a68-a88b-cf875d82bf1e\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526877 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-combined-ca-bundle\") pod \"421b8438-b398-4a68-a88b-cf875d82bf1e\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.526981 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-logs\") pod \"421b8438-b398-4a68-a88b-cf875d82bf1e\" (UID: \"421b8438-b398-4a68-a88b-cf875d82bf1e\") " Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.528653 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-logs" (OuterVolumeSpecName: "logs") pod "421b8438-b398-4a68-a88b-cf875d82bf1e" (UID: "421b8438-b398-4a68-a88b-cf875d82bf1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.532709 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "421b8438-b398-4a68-a88b-cf875d82bf1e" (UID: "421b8438-b398-4a68-a88b-cf875d82bf1e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.533409 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-scripts" (OuterVolumeSpecName: "scripts") pod "421b8438-b398-4a68-a88b-cf875d82bf1e" (UID: "421b8438-b398-4a68-a88b-cf875d82bf1e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.533657 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" (UID: "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.535432 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-logs" (OuterVolumeSpecName: "logs") pod "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" (UID: "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.536973 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-scripts" (OuterVolumeSpecName: "scripts") pod "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" (UID: "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.537354 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-kube-api-access-kf69g" (OuterVolumeSpecName: "kube-api-access-kf69g") pod "421b8438-b398-4a68-a88b-cf875d82bf1e" (UID: "421b8438-b398-4a68-a88b-cf875d82bf1e"). InnerVolumeSpecName "kube-api-access-kf69g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.541115 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-kube-api-access-hmdf4" (OuterVolumeSpecName: "kube-api-access-hmdf4") pod "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" (UID: "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9"). InnerVolumeSpecName "kube-api-access-hmdf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.541816 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-ceph" (OuterVolumeSpecName: "ceph") pod "421b8438-b398-4a68-a88b-cf875d82bf1e" (UID: "421b8438-b398-4a68-a88b-cf875d82bf1e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.542026 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-ceph" (OuterVolumeSpecName: "ceph") pod "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" (UID: "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.564055 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" (UID: "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.566437 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "421b8438-b398-4a68-a88b-cf875d82bf1e" (UID: "421b8438-b398-4a68-a88b-cf875d82bf1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.604415 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-config-data" (OuterVolumeSpecName: "config-data") pod "421b8438-b398-4a68-a88b-cf875d82bf1e" (UID: "421b8438-b398-4a68-a88b-cf875d82bf1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.604462 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-config-data" (OuterVolumeSpecName: "config-data") pod "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" (UID: "2b8d7f12-35c2-47f7-b6a5-d4699b0467c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.611727 4935 generic.go:334] "Generic (PLEG): container finished" podID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerID="c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385" exitCode=0 Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.611828 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9","Type":"ContainerDied","Data":"c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385"} Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.611921 4935 scope.go:117] "RemoveContainer" containerID="c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.611969 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b8d7f12-35c2-47f7-b6a5-d4699b0467c9","Type":"ContainerDied","Data":"6a389b5c8a7e72358b6d68406c69ceb1849df6ae6e78ef162dcd9d7e110e9816"} Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.611997 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.623629 4935 generic.go:334] "Generic (PLEG): container finished" podID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerID="7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b" exitCode=0 Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.623677 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"421b8438-b398-4a68-a88b-cf875d82bf1e","Type":"ContainerDied","Data":"7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b"} Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.623703 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"421b8438-b398-4a68-a88b-cf875d82bf1e","Type":"ContainerDied","Data":"d182e4fc3338243dc6f6313ef64eea424d7dc9204379bec9285ca0895442cea8"} Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.623754 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628515 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628539 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628548 4935 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628557 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmdf4\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-kube-api-access-hmdf4\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628566 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628577 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628584 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628592 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628600 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628608 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf69g\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-kube-api-access-kf69g\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628617 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421b8438-b398-4a68-a88b-cf875d82bf1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628624 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421b8438-b398-4a68-a88b-cf875d82bf1e-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628631 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/421b8438-b398-4a68-a88b-cf875d82bf1e-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.628638 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.656421 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.672415 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.672450 4935 scope.go:117] "RemoveContainer" containerID="9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.684975 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:56:59 crc kubenswrapper[4935]: E1005 08:56:59.688823 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerName="glance-log" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.688839 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerName="glance-log" Oct 05 08:56:59 crc kubenswrapper[4935]: E1005 08:56:59.688870 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerName="glance-httpd" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.688879 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerName="glance-httpd" Oct 05 08:56:59 crc kubenswrapper[4935]: E1005 08:56:59.688907 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerName="glance-log" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.688914 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerName="glance-log" Oct 05 08:56:59 crc kubenswrapper[4935]: E1005 08:56:59.688928 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerName="glance-httpd" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.688934 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerName="glance-httpd" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.689164 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerName="glance-httpd" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.689177 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerName="glance-httpd" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.689188 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" containerName="glance-log" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.689198 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" containerName="glance-log" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.696491 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.700677 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.701351 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.705941 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.723489 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dz7vr" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.732304 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mldb2\" (UniqueName: \"kubernetes.io/projected/59119f30-1499-4892-a9f8-997e38750ded-kube-api-access-mldb2\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.732375 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59119f30-1499-4892-a9f8-997e38750ded-logs\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.732417 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59119f30-1499-4892-a9f8-997e38750ded-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.732824 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/59119f30-1499-4892-a9f8-997e38750ded-ceph\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.732947 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-config-data\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.732975 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.733020 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-scripts\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.771934 4935 scope.go:117] "RemoveContainer" containerID="c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.773992 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:56:59 crc kubenswrapper[4935]: E1005 08:56:59.778091 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385\": container with ID starting with c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385 not found: ID does not exist" containerID="c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.778173 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385"} err="failed to get container status \"c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385\": rpc error: code = NotFound desc = could not find container \"c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385\": container with ID starting with c76ddd5d4bd25a401b700d0cb310246b0f42b7ee7c3e89a1270317229c0ff385 not found: ID does not exist" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.778200 4935 scope.go:117] "RemoveContainer" containerID="9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb" Oct 05 08:56:59 crc kubenswrapper[4935]: E1005 08:56:59.780245 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb\": container with ID starting with 9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb not found: ID does not exist" containerID="9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.780331 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb"} err="failed to get container status \"9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb\": rpc error: code = NotFound desc = could not find container \"9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb\": container with ID starting with 9b17f4e28b560290249c6af72ac9c2559189afa10dd9c6dc01f93946d881d7cb not found: ID does not exist" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.780359 4935 scope.go:117] "RemoveContainer" containerID="7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.786038 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.807297 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.809214 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.814537 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.819248 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.825339 4935 scope.go:117] "RemoveContainer" containerID="62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835180 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mldb2\" (UniqueName: \"kubernetes.io/projected/59119f30-1499-4892-a9f8-997e38750ded-kube-api-access-mldb2\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835315 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59119f30-1499-4892-a9f8-997e38750ded-logs\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835509 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59119f30-1499-4892-a9f8-997e38750ded-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835552 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/59119f30-1499-4892-a9f8-997e38750ded-ceph\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835610 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835691 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835750 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835774 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef490248-2d23-4834-8a80-334916da488f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835821 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef490248-2d23-4834-8a80-334916da488f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835850 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-config-data\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835902 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835940 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef490248-2d23-4834-8a80-334916da488f-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.835990 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmh8d\" (UniqueName: \"kubernetes.io/projected/ef490248-2d23-4834-8a80-334916da488f-kube-api-access-kmh8d\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.836022 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-scripts\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.836090 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59119f30-1499-4892-a9f8-997e38750ded-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.836331 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59119f30-1499-4892-a9f8-997e38750ded-logs\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.840491 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-scripts\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.841266 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.842247 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/59119f30-1499-4892-a9f8-997e38750ded-ceph\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.854283 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59119f30-1499-4892-a9f8-997e38750ded-config-data\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.858197 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mldb2\" (UniqueName: \"kubernetes.io/projected/59119f30-1499-4892-a9f8-997e38750ded-kube-api-access-mldb2\") pod \"glance-default-external-api-0\" (UID: \"59119f30-1499-4892-a9f8-997e38750ded\") " pod="openstack/glance-default-external-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.861527 4935 scope.go:117] "RemoveContainer" containerID="7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b" Oct 05 08:56:59 crc kubenswrapper[4935]: E1005 08:56:59.863454 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b\": container with ID starting with 7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b not found: ID does not exist" containerID="7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.863577 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b"} err="failed to get container status \"7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b\": rpc error: code = NotFound desc = could not find container \"7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b\": container with ID starting with 7dd0904b225e2f82ec43f2049655ed3b39e3557088067c3ba48ce9bb8d25c35b not found: ID does not exist" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.863681 4935 scope.go:117] "RemoveContainer" containerID="62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613" Oct 05 08:56:59 crc kubenswrapper[4935]: E1005 08:56:59.864260 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613\": container with ID starting with 62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613 not found: ID does not exist" containerID="62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.864306 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613"} err="failed to get container status \"62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613\": rpc error: code = NotFound desc = could not find container \"62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613\": container with ID starting with 62745d4098c1629edf7752279c6e2207cdfd357b9405e3d98bf6dfd5e08d6613 not found: ID does not exist" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.938280 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.938418 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.938453 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.938479 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef490248-2d23-4834-8a80-334916da488f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.938508 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef490248-2d23-4834-8a80-334916da488f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.938553 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef490248-2d23-4834-8a80-334916da488f-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.938586 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmh8d\" (UniqueName: \"kubernetes.io/projected/ef490248-2d23-4834-8a80-334916da488f-kube-api-access-kmh8d\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.939332 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef490248-2d23-4834-8a80-334916da488f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.939403 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef490248-2d23-4834-8a80-334916da488f-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.942229 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef490248-2d23-4834-8a80-334916da488f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.943953 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.944741 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.947827 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef490248-2d23-4834-8a80-334916da488f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:56:59 crc kubenswrapper[4935]: I1005 08:56:59.957451 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmh8d\" (UniqueName: \"kubernetes.io/projected/ef490248-2d23-4834-8a80-334916da488f-kube-api-access-kmh8d\") pod \"glance-default-internal-api-0\" (UID: \"ef490248-2d23-4834-8a80-334916da488f\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:57:00 crc kubenswrapper[4935]: I1005 08:57:00.053579 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:57:00 crc kubenswrapper[4935]: I1005 08:57:00.128300 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:00 crc kubenswrapper[4935]: I1005 08:57:00.689793 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:57:00 crc kubenswrapper[4935]: I1005 08:57:00.795754 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b8d7f12-35c2-47f7-b6a5-d4699b0467c9" path="/var/lib/kubelet/pods/2b8d7f12-35c2-47f7-b6a5-d4699b0467c9/volumes" Oct 05 08:57:00 crc kubenswrapper[4935]: I1005 08:57:00.796833 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="421b8438-b398-4a68-a88b-cf875d82bf1e" path="/var/lib/kubelet/pods/421b8438-b398-4a68-a88b-cf875d82bf1e/volumes" Oct 05 08:57:00 crc kubenswrapper[4935]: I1005 08:57:00.797646 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.695363 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bb459497-g9b7r" event={"ID":"2cdeadef-85fc-4477-8352-dc671812251e","Type":"ContainerStarted","Data":"cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38"} Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.697950 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59119f30-1499-4892-a9f8-997e38750ded","Type":"ContainerStarted","Data":"58fcb4da34e6f3bc20ef9a870bfe832b2f4d51f0732c67137a9d63d88774f5b9"} Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.699979 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc74c48f7-mfk2x" event={"ID":"da554f1c-7227-4af1-991d-e9431ba90e68","Type":"ContainerStarted","Data":"fd115b2689f22199d2b01d1014c8a07a5d52cc008e2eb7691a2217f70377d04a"} Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.701595 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef490248-2d23-4834-8a80-334916da488f","Type":"ContainerStarted","Data":"8ef8146dfd8bb9bb45670dc8dba5f3bddcd1c0b4ccbe70c186591f6c6aa6ff9a"} Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.705474 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66575b9d69-5f2nd" event={"ID":"1b87908c-7e5c-402d-80bf-954d6a601f64","Type":"ContainerStarted","Data":"46cc683492e292c823b7c599f0579fe7bc20f5b850569fd0c93f26c9e257ef30"} Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.705544 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66575b9d69-5f2nd" event={"ID":"1b87908c-7e5c-402d-80bf-954d6a601f64","Type":"ContainerStarted","Data":"cd4f2a22cbb1af1115b654828a725a3b9d6b29671cf9dcfd6226bc1d609e5f58"} Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.705779 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66575b9d69-5f2nd" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerName="horizon-log" containerID="cri-o://cd4f2a22cbb1af1115b654828a725a3b9d6b29671cf9dcfd6226bc1d609e5f58" gracePeriod=30 Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.706829 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66575b9d69-5f2nd" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerName="horizon" containerID="cri-o://46cc683492e292c823b7c599f0579fe7bc20f5b850569fd0c93f26c9e257ef30" gracePeriod=30 Oct 05 08:57:05 crc kubenswrapper[4935]: I1005 08:57:05.738458 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66575b9d69-5f2nd" podStartSLOduration=2.365840494 podStartE2EDuration="10.738440113s" podCreationTimestamp="2025-10-05 08:56:55 +0000 UTC" firstStartedPulling="2025-10-05 08:56:56.868689804 +0000 UTC m=+7450.751316264" lastFinishedPulling="2025-10-05 08:57:05.241289363 +0000 UTC m=+7459.123915883" observedRunningTime="2025-10-05 08:57:05.728296174 +0000 UTC m=+7459.610922654" watchObservedRunningTime="2025-10-05 08:57:05.738440113 +0000 UTC m=+7459.621066583" Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.061490 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.717447 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59119f30-1499-4892-a9f8-997e38750ded","Type":"ContainerStarted","Data":"b4e3c0786834ccc5570d132b7b82084b6a528078b58b846a0f4e27daa9901caf"} Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.717740 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59119f30-1499-4892-a9f8-997e38750ded","Type":"ContainerStarted","Data":"b22c0184e1158f2fe3b7faa1d44cd3003b178f8b98e67d61a382c0b031653f50"} Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.720667 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc74c48f7-mfk2x" event={"ID":"da554f1c-7227-4af1-991d-e9431ba90e68","Type":"ContainerStarted","Data":"55e37872e03da5d9ee19ba97994445053b2e1922313d6dba1bde1a2438659b4d"} Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.723169 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef490248-2d23-4834-8a80-334916da488f","Type":"ContainerStarted","Data":"e12eef0c49732dd5d0b6ab8e35b967a51137b224fc7b4bbd8ea1ce6f8ac5609c"} Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.723208 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ef490248-2d23-4834-8a80-334916da488f","Type":"ContainerStarted","Data":"ea89bb19cbfa6e6f27f887c4b5daf6f086f01a00b70da6b01d5bf93bcdaacb4a"} Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.727567 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bb459497-g9b7r" event={"ID":"2cdeadef-85fc-4477-8352-dc671812251e","Type":"ContainerStarted","Data":"4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73"} Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.742820 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.742799842 podStartE2EDuration="7.742799842s" podCreationTimestamp="2025-10-05 08:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:57:06.738333433 +0000 UTC m=+7460.620959883" watchObservedRunningTime="2025-10-05 08:57:06.742799842 +0000 UTC m=+7460.625426312" Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.754497 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-75bb459497-g9b7r" podStartSLOduration=2.705483311 podStartE2EDuration="10.754481111s" podCreationTimestamp="2025-10-05 08:56:56 +0000 UTC" firstStartedPulling="2025-10-05 08:56:57.194544273 +0000 UTC m=+7451.077170733" lastFinishedPulling="2025-10-05 08:57:05.243542033 +0000 UTC m=+7459.126168533" observedRunningTime="2025-10-05 08:57:06.754155433 +0000 UTC m=+7460.636781903" watchObservedRunningTime="2025-10-05 08:57:06.754481111 +0000 UTC m=+7460.637107561" Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.800038 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.800021749 podStartE2EDuration="7.800021749s" podCreationTimestamp="2025-10-05 08:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:57:06.795952731 +0000 UTC m=+7460.678579221" watchObservedRunningTime="2025-10-05 08:57:06.800021749 +0000 UTC m=+7460.682648209" Oct 05 08:57:06 crc kubenswrapper[4935]: I1005 08:57:06.803637 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7bc74c48f7-mfk2x" podStartSLOduration=3.135987783 podStartE2EDuration="11.803631295s" podCreationTimestamp="2025-10-05 08:56:55 +0000 UTC" firstStartedPulling="2025-10-05 08:56:56.562443354 +0000 UTC m=+7450.445069814" lastFinishedPulling="2025-10-05 08:57:05.230086856 +0000 UTC m=+7459.112713326" observedRunningTime="2025-10-05 08:57:06.77553138 +0000 UTC m=+7460.658157850" watchObservedRunningTime="2025-10-05 08:57:06.803631295 +0000 UTC m=+7460.686257755" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.053995 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.054406 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.098243 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.113762 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.128880 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.128958 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.166726 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.175420 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.772548 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.772977 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.772998 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:10 crc kubenswrapper[4935]: I1005 08:57:10.773017 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:13 crc kubenswrapper[4935]: I1005 08:57:13.375733 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:13 crc kubenswrapper[4935]: I1005 08:57:13.379794 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 08:57:13 crc kubenswrapper[4935]: I1005 08:57:13.849462 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 08:57:13 crc kubenswrapper[4935]: I1005 08:57:13.903531 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 08:57:14 crc kubenswrapper[4935]: I1005 08:57:14.289656 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:57:14 crc kubenswrapper[4935]: I1005 08:57:14.289740 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:57:15 crc kubenswrapper[4935]: I1005 08:57:15.935616 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:57:15 crc kubenswrapper[4935]: I1005 08:57:15.936077 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:57:15 crc kubenswrapper[4935]: I1005 08:57:15.937423 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc74c48f7-mfk2x" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.105:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.105:8080: connect: connection refused" Oct 05 08:57:16 crc kubenswrapper[4935]: I1005 08:57:16.713563 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:57:16 crc kubenswrapper[4935]: I1005 08:57:16.715759 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:57:16 crc kubenswrapper[4935]: I1005 08:57:16.721049 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-75bb459497-g9b7r" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 05 08:57:28 crc kubenswrapper[4935]: I1005 08:57:28.050412 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:57:28 crc kubenswrapper[4935]: I1005 08:57:28.460414 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:57:29 crc kubenswrapper[4935]: I1005 08:57:29.709047 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:57:30 crc kubenswrapper[4935]: I1005 08:57:30.126351 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:57:30 crc kubenswrapper[4935]: I1005 08:57:30.200614 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bc74c48f7-mfk2x"] Oct 05 08:57:30 crc kubenswrapper[4935]: I1005 08:57:30.201065 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bc74c48f7-mfk2x" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon-log" containerID="cri-o://fd115b2689f22199d2b01d1014c8a07a5d52cc008e2eb7691a2217f70377d04a" gracePeriod=30 Oct 05 08:57:30 crc kubenswrapper[4935]: I1005 08:57:30.201194 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bc74c48f7-mfk2x" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon" containerID="cri-o://55e37872e03da5d9ee19ba97994445053b2e1922313d6dba1bde1a2438659b4d" gracePeriod=30 Oct 05 08:57:34 crc kubenswrapper[4935]: I1005 08:57:34.023460 4935 generic.go:334] "Generic (PLEG): container finished" podID="da554f1c-7227-4af1-991d-e9431ba90e68" containerID="55e37872e03da5d9ee19ba97994445053b2e1922313d6dba1bde1a2438659b4d" exitCode=0 Oct 05 08:57:34 crc kubenswrapper[4935]: I1005 08:57:34.023515 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc74c48f7-mfk2x" event={"ID":"da554f1c-7227-4af1-991d-e9431ba90e68","Type":"ContainerDied","Data":"55e37872e03da5d9ee19ba97994445053b2e1922313d6dba1bde1a2438659b4d"} Oct 05 08:57:35 crc kubenswrapper[4935]: I1005 08:57:35.937063 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bc74c48f7-mfk2x" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.105:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.105:8080: connect: connection refused" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.057988 4935 generic.go:334] "Generic (PLEG): container finished" podID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerID="46cc683492e292c823b7c599f0579fe7bc20f5b850569fd0c93f26c9e257ef30" exitCode=137 Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.058054 4935 generic.go:334] "Generic (PLEG): container finished" podID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerID="cd4f2a22cbb1af1115b654828a725a3b9d6b29671cf9dcfd6226bc1d609e5f58" exitCode=137 Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.058058 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66575b9d69-5f2nd" event={"ID":"1b87908c-7e5c-402d-80bf-954d6a601f64","Type":"ContainerDied","Data":"46cc683492e292c823b7c599f0579fe7bc20f5b850569fd0c93f26c9e257ef30"} Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.058135 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66575b9d69-5f2nd" event={"ID":"1b87908c-7e5c-402d-80bf-954d6a601f64","Type":"ContainerDied","Data":"cd4f2a22cbb1af1115b654828a725a3b9d6b29671cf9dcfd6226bc1d609e5f58"} Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.254454 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.274192 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-scripts\") pod \"1b87908c-7e5c-402d-80bf-954d6a601f64\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.274965 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b87908c-7e5c-402d-80bf-954d6a601f64-horizon-secret-key\") pod \"1b87908c-7e5c-402d-80bf-954d6a601f64\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.275121 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-config-data\") pod \"1b87908c-7e5c-402d-80bf-954d6a601f64\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.275224 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/1b87908c-7e5c-402d-80bf-954d6a601f64-kube-api-access-pfrsf\") pod \"1b87908c-7e5c-402d-80bf-954d6a601f64\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.275314 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b87908c-7e5c-402d-80bf-954d6a601f64-logs\") pod \"1b87908c-7e5c-402d-80bf-954d6a601f64\" (UID: \"1b87908c-7e5c-402d-80bf-954d6a601f64\") " Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.276147 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b87908c-7e5c-402d-80bf-954d6a601f64-logs" (OuterVolumeSpecName: "logs") pod "1b87908c-7e5c-402d-80bf-954d6a601f64" (UID: "1b87908c-7e5c-402d-80bf-954d6a601f64"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.277048 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b87908c-7e5c-402d-80bf-954d6a601f64-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.287767 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b87908c-7e5c-402d-80bf-954d6a601f64-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1b87908c-7e5c-402d-80bf-954d6a601f64" (UID: "1b87908c-7e5c-402d-80bf-954d6a601f64"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.294351 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b87908c-7e5c-402d-80bf-954d6a601f64-kube-api-access-pfrsf" (OuterVolumeSpecName: "kube-api-access-pfrsf") pod "1b87908c-7e5c-402d-80bf-954d6a601f64" (UID: "1b87908c-7e5c-402d-80bf-954d6a601f64"). InnerVolumeSpecName "kube-api-access-pfrsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.332824 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-config-data" (OuterVolumeSpecName: "config-data") pod "1b87908c-7e5c-402d-80bf-954d6a601f64" (UID: "1b87908c-7e5c-402d-80bf-954d6a601f64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.333605 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-scripts" (OuterVolumeSpecName: "scripts") pod "1b87908c-7e5c-402d-80bf-954d6a601f64" (UID: "1b87908c-7e5c-402d-80bf-954d6a601f64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.378576 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.378625 4935 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b87908c-7e5c-402d-80bf-954d6a601f64-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.378638 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b87908c-7e5c-402d-80bf-954d6a601f64-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:57:36 crc kubenswrapper[4935]: I1005 08:57:36.378650 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/1b87908c-7e5c-402d-80bf-954d6a601f64-kube-api-access-pfrsf\") on node \"crc\" DevicePath \"\"" Oct 05 08:57:37 crc kubenswrapper[4935]: I1005 08:57:37.084981 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66575b9d69-5f2nd" event={"ID":"1b87908c-7e5c-402d-80bf-954d6a601f64","Type":"ContainerDied","Data":"d820bdc88b26636bbc22a8bf227792c4ccb1eade1f733577e35a729c0cacb017"} Oct 05 08:57:37 crc kubenswrapper[4935]: I1005 08:57:37.085045 4935 scope.go:117] "RemoveContainer" containerID="46cc683492e292c823b7c599f0579fe7bc20f5b850569fd0c93f26c9e257ef30" Oct 05 08:57:37 crc kubenswrapper[4935]: I1005 08:57:37.085140 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66575b9d69-5f2nd" Oct 05 08:57:37 crc kubenswrapper[4935]: I1005 08:57:37.126431 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66575b9d69-5f2nd"] Oct 05 08:57:37 crc kubenswrapper[4935]: I1005 08:57:37.137316 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66575b9d69-5f2nd"] Oct 05 08:57:37 crc kubenswrapper[4935]: I1005 08:57:37.294673 4935 scope.go:117] "RemoveContainer" containerID="cd4f2a22cbb1af1115b654828a725a3b9d6b29671cf9dcfd6226bc1d609e5f58" Oct 05 08:57:38 crc kubenswrapper[4935]: I1005 08:57:38.793487 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" path="/var/lib/kubelet/pods/1b87908c-7e5c-402d-80bf-954d6a601f64/volumes" Oct 05 08:57:44 crc kubenswrapper[4935]: I1005 08:57:44.289442 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:57:44 crc kubenswrapper[4935]: I1005 08:57:44.290398 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:57:44 crc kubenswrapper[4935]: I1005 08:57:44.290475 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 08:57:44 crc kubenswrapper[4935]: I1005 08:57:44.291760 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4f78b54294a258ac737523e93c508737226a48f0df445e1a69f2784b25cc51ab"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:57:44 crc kubenswrapper[4935]: I1005 08:57:44.291871 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://4f78b54294a258ac737523e93c508737226a48f0df445e1a69f2784b25cc51ab" gracePeriod=600 Oct 05 08:57:45 crc kubenswrapper[4935]: I1005 08:57:45.063986 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-677g4"] Oct 05 08:57:45 crc kubenswrapper[4935]: I1005 08:57:45.076921 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-677g4"] Oct 05 08:57:45 crc kubenswrapper[4935]: I1005 08:57:45.189351 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="4f78b54294a258ac737523e93c508737226a48f0df445e1a69f2784b25cc51ab" exitCode=0 Oct 05 08:57:45 crc kubenswrapper[4935]: I1005 08:57:45.189412 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"4f78b54294a258ac737523e93c508737226a48f0df445e1a69f2784b25cc51ab"} Oct 05 08:57:45 crc kubenswrapper[4935]: I1005 08:57:45.189489 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652"} Oct 05 08:57:45 crc kubenswrapper[4935]: I1005 08:57:45.189521 4935 scope.go:117] "RemoveContainer" containerID="05d02a6a44d6e79766059ae3f1c0e85106053066a435535d653a6c24cea18899" Oct 05 08:57:45 crc kubenswrapper[4935]: I1005 08:57:45.936199 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bc74c48f7-mfk2x" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.105:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.105:8080: connect: connection refused" Oct 05 08:57:46 crc kubenswrapper[4935]: I1005 08:57:46.795817 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a2c18d3-0539-408d-acd7-227d032f1787" path="/var/lib/kubelet/pods/6a2c18d3-0539-408d-acd7-227d032f1787/volumes" Oct 05 08:57:55 crc kubenswrapper[4935]: I1005 08:57:55.029272 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7798-account-create-xhjk2"] Oct 05 08:57:55 crc kubenswrapper[4935]: I1005 08:57:55.041239 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7798-account-create-xhjk2"] Oct 05 08:57:55 crc kubenswrapper[4935]: I1005 08:57:55.936670 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bc74c48f7-mfk2x" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.105:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.105:8080: connect: connection refused" Oct 05 08:57:55 crc kubenswrapper[4935]: I1005 08:57:55.937052 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:57:56 crc kubenswrapper[4935]: I1005 08:57:56.791561 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f0f894-2eef-4b8a-9e38-3e71e7c0283a" path="/var/lib/kubelet/pods/00f0f894-2eef-4b8a-9e38-3e71e7c0283a/volumes" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.309375 4935 scope.go:117] "RemoveContainer" containerID="8779c26fab3cd3f7b5820571169be88c752fd145ae4409ebd8b6fb049b5375e1" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.379190 4935 generic.go:334] "Generic (PLEG): container finished" podID="da554f1c-7227-4af1-991d-e9431ba90e68" containerID="fd115b2689f22199d2b01d1014c8a07a5d52cc008e2eb7691a2217f70377d04a" exitCode=137 Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.379242 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc74c48f7-mfk2x" event={"ID":"da554f1c-7227-4af1-991d-e9431ba90e68","Type":"ContainerDied","Data":"fd115b2689f22199d2b01d1014c8a07a5d52cc008e2eb7691a2217f70377d04a"} Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.405878 4935 scope.go:117] "RemoveContainer" containerID="d454166d0eb85993d4ae1fd650b5c910c14f648dca66d693b01e551a36d34764" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.686979 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.810379 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-scripts\") pod \"da554f1c-7227-4af1-991d-e9431ba90e68\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.810426 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-config-data\") pod \"da554f1c-7227-4af1-991d-e9431ba90e68\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.810496 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da554f1c-7227-4af1-991d-e9431ba90e68-logs\") pod \"da554f1c-7227-4af1-991d-e9431ba90e68\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.810543 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4zwd\" (UniqueName: \"kubernetes.io/projected/da554f1c-7227-4af1-991d-e9431ba90e68-kube-api-access-c4zwd\") pod \"da554f1c-7227-4af1-991d-e9431ba90e68\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.810636 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da554f1c-7227-4af1-991d-e9431ba90e68-horizon-secret-key\") pod \"da554f1c-7227-4af1-991d-e9431ba90e68\" (UID: \"da554f1c-7227-4af1-991d-e9431ba90e68\") " Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.811215 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da554f1c-7227-4af1-991d-e9431ba90e68-logs" (OuterVolumeSpecName: "logs") pod "da554f1c-7227-4af1-991d-e9431ba90e68" (UID: "da554f1c-7227-4af1-991d-e9431ba90e68"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.811324 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da554f1c-7227-4af1-991d-e9431ba90e68-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.816209 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da554f1c-7227-4af1-991d-e9431ba90e68-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "da554f1c-7227-4af1-991d-e9431ba90e68" (UID: "da554f1c-7227-4af1-991d-e9431ba90e68"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.817392 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da554f1c-7227-4af1-991d-e9431ba90e68-kube-api-access-c4zwd" (OuterVolumeSpecName: "kube-api-access-c4zwd") pod "da554f1c-7227-4af1-991d-e9431ba90e68" (UID: "da554f1c-7227-4af1-991d-e9431ba90e68"). InnerVolumeSpecName "kube-api-access-c4zwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.843281 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-scripts" (OuterVolumeSpecName: "scripts") pod "da554f1c-7227-4af1-991d-e9431ba90e68" (UID: "da554f1c-7227-4af1-991d-e9431ba90e68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.850709 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-config-data" (OuterVolumeSpecName: "config-data") pod "da554f1c-7227-4af1-991d-e9431ba90e68" (UID: "da554f1c-7227-4af1-991d-e9431ba90e68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.912511 4935 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da554f1c-7227-4af1-991d-e9431ba90e68-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.912546 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.912555 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da554f1c-7227-4af1-991d-e9431ba90e68-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:00 crc kubenswrapper[4935]: I1005 08:58:00.912564 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4zwd\" (UniqueName: \"kubernetes.io/projected/da554f1c-7227-4af1-991d-e9431ba90e68-kube-api-access-c4zwd\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:01 crc kubenswrapper[4935]: I1005 08:58:01.389749 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc74c48f7-mfk2x" event={"ID":"da554f1c-7227-4af1-991d-e9431ba90e68","Type":"ContainerDied","Data":"22e87679d902ea723abac1c6b63278c1987feb3c6923a1a8b79ed91dbecd3c16"} Oct 05 08:58:01 crc kubenswrapper[4935]: I1005 08:58:01.389918 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc74c48f7-mfk2x" Oct 05 08:58:01 crc kubenswrapper[4935]: I1005 08:58:01.390113 4935 scope.go:117] "RemoveContainer" containerID="55e37872e03da5d9ee19ba97994445053b2e1922313d6dba1bde1a2438659b4d" Oct 05 08:58:01 crc kubenswrapper[4935]: I1005 08:58:01.427863 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bc74c48f7-mfk2x"] Oct 05 08:58:01 crc kubenswrapper[4935]: I1005 08:58:01.449186 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bc74c48f7-mfk2x"] Oct 05 08:58:01 crc kubenswrapper[4935]: I1005 08:58:01.574299 4935 scope.go:117] "RemoveContainer" containerID="fd115b2689f22199d2b01d1014c8a07a5d52cc008e2eb7691a2217f70377d04a" Oct 05 08:58:02 crc kubenswrapper[4935]: I1005 08:58:02.793790 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" path="/var/lib/kubelet/pods/da554f1c-7227-4af1-991d-e9431ba90e68/volumes" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.586483 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67b8b7d97c-2hshb"] Oct 05 08:58:04 crc kubenswrapper[4935]: E1005 08:58:04.587601 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerName="horizon-log" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.587623 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerName="horizon-log" Oct 05 08:58:04 crc kubenswrapper[4935]: E1005 08:58:04.587652 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.587662 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon" Oct 05 08:58:04 crc kubenswrapper[4935]: E1005 08:58:04.587680 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerName="horizon" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.587690 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerName="horizon" Oct 05 08:58:04 crc kubenswrapper[4935]: E1005 08:58:04.587712 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon-log" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.587720 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon-log" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.587999 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerName="horizon-log" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.588011 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.588042 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b87908c-7e5c-402d-80bf-954d6a601f64" containerName="horizon" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.588067 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="da554f1c-7227-4af1-991d-e9431ba90e68" containerName="horizon-log" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.589589 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.599554 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67b8b7d97c-2hshb"] Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.696500 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc66c\" (UniqueName: \"kubernetes.io/projected/c15807dd-5d29-4c60-a99e-c80dd38c0b27-kube-api-access-vc66c\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.696576 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c15807dd-5d29-4c60-a99e-c80dd38c0b27-config-data\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.696613 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c15807dd-5d29-4c60-a99e-c80dd38c0b27-scripts\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.696640 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c15807dd-5d29-4c60-a99e-c80dd38c0b27-horizon-secret-key\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.696728 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c15807dd-5d29-4c60-a99e-c80dd38c0b27-logs\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.798216 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc66c\" (UniqueName: \"kubernetes.io/projected/c15807dd-5d29-4c60-a99e-c80dd38c0b27-kube-api-access-vc66c\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.798611 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c15807dd-5d29-4c60-a99e-c80dd38c0b27-config-data\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.798641 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c15807dd-5d29-4c60-a99e-c80dd38c0b27-scripts\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.798657 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c15807dd-5d29-4c60-a99e-c80dd38c0b27-horizon-secret-key\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.798724 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c15807dd-5d29-4c60-a99e-c80dd38c0b27-logs\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.799256 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c15807dd-5d29-4c60-a99e-c80dd38c0b27-logs\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.799774 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c15807dd-5d29-4c60-a99e-c80dd38c0b27-scripts\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.800237 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c15807dd-5d29-4c60-a99e-c80dd38c0b27-config-data\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.806082 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c15807dd-5d29-4c60-a99e-c80dd38c0b27-horizon-secret-key\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.825224 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc66c\" (UniqueName: \"kubernetes.io/projected/c15807dd-5d29-4c60-a99e-c80dd38c0b27-kube-api-access-vc66c\") pod \"horizon-67b8b7d97c-2hshb\" (UID: \"c15807dd-5d29-4c60-a99e-c80dd38c0b27\") " pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:04 crc kubenswrapper[4935]: I1005 08:58:04.956272 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:05 crc kubenswrapper[4935]: I1005 08:58:05.493192 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67b8b7d97c-2hshb"] Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.213927 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-m4w8v"] Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.215478 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-m4w8v" Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.228214 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-m4w8v"] Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.336135 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsv6k\" (UniqueName: \"kubernetes.io/projected/cbd77f57-fc8f-48d7-afcb-37e67611049f-kube-api-access-zsv6k\") pod \"heat-db-create-m4w8v\" (UID: \"cbd77f57-fc8f-48d7-afcb-37e67611049f\") " pod="openstack/heat-db-create-m4w8v" Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.438916 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsv6k\" (UniqueName: \"kubernetes.io/projected/cbd77f57-fc8f-48d7-afcb-37e67611049f-kube-api-access-zsv6k\") pod \"heat-db-create-m4w8v\" (UID: \"cbd77f57-fc8f-48d7-afcb-37e67611049f\") " pod="openstack/heat-db-create-m4w8v" Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.464748 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b8b7d97c-2hshb" event={"ID":"c15807dd-5d29-4c60-a99e-c80dd38c0b27","Type":"ContainerStarted","Data":"adf7499d09b6b62a705ec71ab1c533a6ab69c1f44caeedc72dc47ed5de694416"} Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.464815 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b8b7d97c-2hshb" event={"ID":"c15807dd-5d29-4c60-a99e-c80dd38c0b27","Type":"ContainerStarted","Data":"ecb7a8c539ff20c7d3b4f4b1e4498ecb5184c52a478c616d84249321d2fb573b"} Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.464826 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67b8b7d97c-2hshb" event={"ID":"c15807dd-5d29-4c60-a99e-c80dd38c0b27","Type":"ContainerStarted","Data":"f58e68e86374fabf790ed9aa82793e2f8d38fd73f8e533f9da85a442bbdbff38"} Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.471309 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsv6k\" (UniqueName: \"kubernetes.io/projected/cbd77f57-fc8f-48d7-afcb-37e67611049f-kube-api-access-zsv6k\") pod \"heat-db-create-m4w8v\" (UID: \"cbd77f57-fc8f-48d7-afcb-37e67611049f\") " pod="openstack/heat-db-create-m4w8v" Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.496329 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67b8b7d97c-2hshb" podStartSLOduration=2.496307797 podStartE2EDuration="2.496307797s" podCreationTimestamp="2025-10-05 08:58:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:58:06.495087605 +0000 UTC m=+7520.377714105" watchObservedRunningTime="2025-10-05 08:58:06.496307797 +0000 UTC m=+7520.378934257" Oct 05 08:58:06 crc kubenswrapper[4935]: I1005 08:58:06.547272 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-m4w8v" Oct 05 08:58:07 crc kubenswrapper[4935]: I1005 08:58:07.055094 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-r2sf9"] Oct 05 08:58:07 crc kubenswrapper[4935]: I1005 08:58:07.066146 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-r2sf9"] Oct 05 08:58:07 crc kubenswrapper[4935]: I1005 08:58:07.099092 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-m4w8v"] Oct 05 08:58:07 crc kubenswrapper[4935]: I1005 08:58:07.478967 4935 generic.go:334] "Generic (PLEG): container finished" podID="cbd77f57-fc8f-48d7-afcb-37e67611049f" containerID="3d59c116227bf71be70d0fd62ba7516148990b16ced050da9a81d91e9e641ab7" exitCode=0 Oct 05 08:58:07 crc kubenswrapper[4935]: I1005 08:58:07.479026 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-m4w8v" event={"ID":"cbd77f57-fc8f-48d7-afcb-37e67611049f","Type":"ContainerDied","Data":"3d59c116227bf71be70d0fd62ba7516148990b16ced050da9a81d91e9e641ab7"} Oct 05 08:58:07 crc kubenswrapper[4935]: I1005 08:58:07.479487 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-m4w8v" event={"ID":"cbd77f57-fc8f-48d7-afcb-37e67611049f","Type":"ContainerStarted","Data":"5b4d8426c083140477cef9a13da03d1b861492f63c0349075d6a0ed8d1ed30f1"} Oct 05 08:58:08 crc kubenswrapper[4935]: I1005 08:58:08.799279 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e2eb559-8cea-4191-9f10-5b564959ac9b" path="/var/lib/kubelet/pods/7e2eb559-8cea-4191-9f10-5b564959ac9b/volumes" Oct 05 08:58:08 crc kubenswrapper[4935]: I1005 08:58:08.968672 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-m4w8v" Oct 05 08:58:09 crc kubenswrapper[4935]: I1005 08:58:09.100045 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsv6k\" (UniqueName: \"kubernetes.io/projected/cbd77f57-fc8f-48d7-afcb-37e67611049f-kube-api-access-zsv6k\") pod \"cbd77f57-fc8f-48d7-afcb-37e67611049f\" (UID: \"cbd77f57-fc8f-48d7-afcb-37e67611049f\") " Oct 05 08:58:09 crc kubenswrapper[4935]: I1005 08:58:09.108814 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbd77f57-fc8f-48d7-afcb-37e67611049f-kube-api-access-zsv6k" (OuterVolumeSpecName: "kube-api-access-zsv6k") pod "cbd77f57-fc8f-48d7-afcb-37e67611049f" (UID: "cbd77f57-fc8f-48d7-afcb-37e67611049f"). InnerVolumeSpecName "kube-api-access-zsv6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:58:09 crc kubenswrapper[4935]: I1005 08:58:09.208814 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsv6k\" (UniqueName: \"kubernetes.io/projected/cbd77f57-fc8f-48d7-afcb-37e67611049f-kube-api-access-zsv6k\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:09 crc kubenswrapper[4935]: I1005 08:58:09.502006 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-m4w8v" event={"ID":"cbd77f57-fc8f-48d7-afcb-37e67611049f","Type":"ContainerDied","Data":"5b4d8426c083140477cef9a13da03d1b861492f63c0349075d6a0ed8d1ed30f1"} Oct 05 08:58:09 crc kubenswrapper[4935]: I1005 08:58:09.502046 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-m4w8v" Oct 05 08:58:09 crc kubenswrapper[4935]: I1005 08:58:09.502048 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b4d8426c083140477cef9a13da03d1b861492f63c0349075d6a0ed8d1ed30f1" Oct 05 08:58:14 crc kubenswrapper[4935]: I1005 08:58:14.957090 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:14 crc kubenswrapper[4935]: I1005 08:58:14.957778 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.236212 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-6d32-account-create-9qkt6"] Oct 05 08:58:16 crc kubenswrapper[4935]: E1005 08:58:16.236723 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd77f57-fc8f-48d7-afcb-37e67611049f" containerName="mariadb-database-create" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.236743 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd77f57-fc8f-48d7-afcb-37e67611049f" containerName="mariadb-database-create" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.237021 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbd77f57-fc8f-48d7-afcb-37e67611049f" containerName="mariadb-database-create" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.237797 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6d32-account-create-9qkt6" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.240935 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.250124 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-6d32-account-create-9qkt6"] Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.372934 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njxdj\" (UniqueName: \"kubernetes.io/projected/adf81e84-5745-4f16-acc2-3ee390046548-kube-api-access-njxdj\") pod \"heat-6d32-account-create-9qkt6\" (UID: \"adf81e84-5745-4f16-acc2-3ee390046548\") " pod="openstack/heat-6d32-account-create-9qkt6" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.476323 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njxdj\" (UniqueName: \"kubernetes.io/projected/adf81e84-5745-4f16-acc2-3ee390046548-kube-api-access-njxdj\") pod \"heat-6d32-account-create-9qkt6\" (UID: \"adf81e84-5745-4f16-acc2-3ee390046548\") " pod="openstack/heat-6d32-account-create-9qkt6" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.498477 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njxdj\" (UniqueName: \"kubernetes.io/projected/adf81e84-5745-4f16-acc2-3ee390046548-kube-api-access-njxdj\") pod \"heat-6d32-account-create-9qkt6\" (UID: \"adf81e84-5745-4f16-acc2-3ee390046548\") " pod="openstack/heat-6d32-account-create-9qkt6" Oct 05 08:58:16 crc kubenswrapper[4935]: I1005 08:58:16.575601 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6d32-account-create-9qkt6" Oct 05 08:58:17 crc kubenswrapper[4935]: I1005 08:58:17.123720 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-6d32-account-create-9qkt6"] Oct 05 08:58:17 crc kubenswrapper[4935]: W1005 08:58:17.128267 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadf81e84_5745_4f16_acc2_3ee390046548.slice/crio-03603529568d94ea49e91bb1904e30bc37dbf9cc28ce47ae21e46ddcfcf60a89 WatchSource:0}: Error finding container 03603529568d94ea49e91bb1904e30bc37dbf9cc28ce47ae21e46ddcfcf60a89: Status 404 returned error can't find the container with id 03603529568d94ea49e91bb1904e30bc37dbf9cc28ce47ae21e46ddcfcf60a89 Oct 05 08:58:17 crc kubenswrapper[4935]: I1005 08:58:17.575654 4935 generic.go:334] "Generic (PLEG): container finished" podID="adf81e84-5745-4f16-acc2-3ee390046548" containerID="6b176d45361a51321576e1b653290714ee40f860bafe31cd3cf256bb538ac401" exitCode=0 Oct 05 08:58:17 crc kubenswrapper[4935]: I1005 08:58:17.575699 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6d32-account-create-9qkt6" event={"ID":"adf81e84-5745-4f16-acc2-3ee390046548","Type":"ContainerDied","Data":"6b176d45361a51321576e1b653290714ee40f860bafe31cd3cf256bb538ac401"} Oct 05 08:58:17 crc kubenswrapper[4935]: I1005 08:58:17.576116 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6d32-account-create-9qkt6" event={"ID":"adf81e84-5745-4f16-acc2-3ee390046548","Type":"ContainerStarted","Data":"03603529568d94ea49e91bb1904e30bc37dbf9cc28ce47ae21e46ddcfcf60a89"} Oct 05 08:58:18 crc kubenswrapper[4935]: I1005 08:58:18.957356 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6d32-account-create-9qkt6" Oct 05 08:58:19 crc kubenswrapper[4935]: I1005 08:58:19.029835 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njxdj\" (UniqueName: \"kubernetes.io/projected/adf81e84-5745-4f16-acc2-3ee390046548-kube-api-access-njxdj\") pod \"adf81e84-5745-4f16-acc2-3ee390046548\" (UID: \"adf81e84-5745-4f16-acc2-3ee390046548\") " Oct 05 08:58:19 crc kubenswrapper[4935]: I1005 08:58:19.035176 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf81e84-5745-4f16-acc2-3ee390046548-kube-api-access-njxdj" (OuterVolumeSpecName: "kube-api-access-njxdj") pod "adf81e84-5745-4f16-acc2-3ee390046548" (UID: "adf81e84-5745-4f16-acc2-3ee390046548"). InnerVolumeSpecName "kube-api-access-njxdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:58:19 crc kubenswrapper[4935]: I1005 08:58:19.132052 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njxdj\" (UniqueName: \"kubernetes.io/projected/adf81e84-5745-4f16-acc2-3ee390046548-kube-api-access-njxdj\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:19 crc kubenswrapper[4935]: I1005 08:58:19.600068 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6d32-account-create-9qkt6" event={"ID":"adf81e84-5745-4f16-acc2-3ee390046548","Type":"ContainerDied","Data":"03603529568d94ea49e91bb1904e30bc37dbf9cc28ce47ae21e46ddcfcf60a89"} Oct 05 08:58:19 crc kubenswrapper[4935]: I1005 08:58:19.600119 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03603529568d94ea49e91bb1904e30bc37dbf9cc28ce47ae21e46ddcfcf60a89" Oct 05 08:58:19 crc kubenswrapper[4935]: I1005 08:58:19.600126 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6d32-account-create-9qkt6" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.313074 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-9pd72"] Oct 05 08:58:21 crc kubenswrapper[4935]: E1005 08:58:21.313739 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf81e84-5745-4f16-acc2-3ee390046548" containerName="mariadb-account-create" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.313759 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf81e84-5745-4f16-acc2-3ee390046548" containerName="mariadb-account-create" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.314154 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf81e84-5745-4f16-acc2-3ee390046548" containerName="mariadb-account-create" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.317044 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.323547 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-phcjz" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.329246 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.331781 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-9pd72"] Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.395434 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9cp8\" (UniqueName: \"kubernetes.io/projected/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-kube-api-access-p9cp8\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.395740 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-config-data\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.395883 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-combined-ca-bundle\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.497804 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9cp8\" (UniqueName: \"kubernetes.io/projected/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-kube-api-access-p9cp8\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.497934 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-config-data\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.498114 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-combined-ca-bundle\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.508051 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-combined-ca-bundle\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.513953 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-config-data\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.518705 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9cp8\" (UniqueName: \"kubernetes.io/projected/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-kube-api-access-p9cp8\") pod \"heat-db-sync-9pd72\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:21 crc kubenswrapper[4935]: I1005 08:58:21.675149 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:22 crc kubenswrapper[4935]: I1005 08:58:22.209051 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-9pd72"] Oct 05 08:58:22 crc kubenswrapper[4935]: I1005 08:58:22.215728 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:58:22 crc kubenswrapper[4935]: I1005 08:58:22.629677 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9pd72" event={"ID":"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8","Type":"ContainerStarted","Data":"5040f1a0bc02f90899d6dc37d9fd18dd8b6669548339417477fb4d010db3a2fe"} Oct 05 08:58:26 crc kubenswrapper[4935]: I1005 08:58:26.619540 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:28 crc kubenswrapper[4935]: I1005 08:58:28.301664 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-67b8b7d97c-2hshb" Oct 05 08:58:28 crc kubenswrapper[4935]: I1005 08:58:28.402294 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75bb459497-g9b7r"] Oct 05 08:58:28 crc kubenswrapper[4935]: I1005 08:58:28.402523 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-75bb459497-g9b7r" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon-log" containerID="cri-o://cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38" gracePeriod=30 Oct 05 08:58:28 crc kubenswrapper[4935]: I1005 08:58:28.402967 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-75bb459497-g9b7r" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon" containerID="cri-o://4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73" gracePeriod=30 Oct 05 08:58:30 crc kubenswrapper[4935]: I1005 08:58:30.703876 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9pd72" event={"ID":"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8","Type":"ContainerStarted","Data":"c7e854e4ff1d796d9c50575c54bc8d4800e56ae66443626cb9b95683116536b6"} Oct 05 08:58:30 crc kubenswrapper[4935]: I1005 08:58:30.723018 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-9pd72" podStartSLOduration=1.9168179090000002 podStartE2EDuration="9.722996871s" podCreationTimestamp="2025-10-05 08:58:21 +0000 UTC" firstStartedPulling="2025-10-05 08:58:22.215474264 +0000 UTC m=+7536.098100724" lastFinishedPulling="2025-10-05 08:58:30.021653236 +0000 UTC m=+7543.904279686" observedRunningTime="2025-10-05 08:58:30.72144319 +0000 UTC m=+7544.604069690" watchObservedRunningTime="2025-10-05 08:58:30.722996871 +0000 UTC m=+7544.605623351" Oct 05 08:58:31 crc kubenswrapper[4935]: I1005 08:58:31.030927 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-fmvtl"] Oct 05 08:58:31 crc kubenswrapper[4935]: I1005 08:58:31.039453 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-fmvtl"] Oct 05 08:58:31 crc kubenswrapper[4935]: I1005 08:58:31.713630 4935 generic.go:334] "Generic (PLEG): container finished" podID="2cdeadef-85fc-4477-8352-dc671812251e" containerID="4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73" exitCode=0 Oct 05 08:58:31 crc kubenswrapper[4935]: I1005 08:58:31.713958 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bb459497-g9b7r" event={"ID":"2cdeadef-85fc-4477-8352-dc671812251e","Type":"ContainerDied","Data":"4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73"} Oct 05 08:58:32 crc kubenswrapper[4935]: I1005 08:58:32.734950 4935 generic.go:334] "Generic (PLEG): container finished" podID="9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8" containerID="c7e854e4ff1d796d9c50575c54bc8d4800e56ae66443626cb9b95683116536b6" exitCode=0 Oct 05 08:58:32 crc kubenswrapper[4935]: I1005 08:58:32.735073 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9pd72" event={"ID":"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8","Type":"ContainerDied","Data":"c7e854e4ff1d796d9c50575c54bc8d4800e56ae66443626cb9b95683116536b6"} Oct 05 08:58:32 crc kubenswrapper[4935]: I1005 08:58:32.789882 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f7d891-b3c6-4649-9596-f4b3132b47e0" path="/var/lib/kubelet/pods/77f7d891-b3c6-4649-9596-f4b3132b47e0/volumes" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.168775 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.374953 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-config-data\") pod \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.375044 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-combined-ca-bundle\") pod \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.375158 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9cp8\" (UniqueName: \"kubernetes.io/projected/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-kube-api-access-p9cp8\") pod \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\" (UID: \"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8\") " Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.382236 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-kube-api-access-p9cp8" (OuterVolumeSpecName: "kube-api-access-p9cp8") pod "9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8" (UID: "9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8"). InnerVolumeSpecName "kube-api-access-p9cp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.408783 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8" (UID: "9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.443737 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-config-data" (OuterVolumeSpecName: "config-data") pod "9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8" (UID: "9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.477164 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.477204 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.477221 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9cp8\" (UniqueName: \"kubernetes.io/projected/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8-kube-api-access-p9cp8\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.762574 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9pd72" event={"ID":"9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8","Type":"ContainerDied","Data":"5040f1a0bc02f90899d6dc37d9fd18dd8b6669548339417477fb4d010db3a2fe"} Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.762644 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9pd72" Oct 05 08:58:34 crc kubenswrapper[4935]: I1005 08:58:34.762647 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5040f1a0bc02f90899d6dc37d9fd18dd8b6669548339417477fb4d010db3a2fe" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.213787 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-59d5d97855-8b85w"] Oct 05 08:58:36 crc kubenswrapper[4935]: E1005 08:58:36.214943 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8" containerName="heat-db-sync" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.214963 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8" containerName="heat-db-sync" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.215218 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8" containerName="heat-db-sync" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.216028 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.222261 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.222671 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.222829 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-phcjz" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.298759 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-54d757dd4f-264jb"] Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.300834 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.303294 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.339167 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-59d5d97855-8b85w"] Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.368139 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-54d757dd4f-264jb"] Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.391423 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-59984465f-5qgvh"] Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.393062 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.395254 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.420408 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-config-data\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.420579 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgf7k\" (UniqueName: \"kubernetes.io/projected/f3cb9fa5-3711-4f0a-830b-e2d210caada6-kube-api-access-xgf7k\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.420612 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-config-data-custom\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.420654 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-combined-ca-bundle\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.420713 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-combined-ca-bundle\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.420790 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlwhb\" (UniqueName: \"kubernetes.io/projected/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-kube-api-access-wlwhb\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.420939 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-config-data-custom\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.420965 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-config-data\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.429737 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-59984465f-5qgvh"] Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.524994 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-config-data\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525098 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47kgv\" (UniqueName: \"kubernetes.io/projected/dfeb90b7-9abc-4f02-8a73-2c6df751c902-kube-api-access-47kgv\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525137 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-combined-ca-bundle\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525163 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-config-data-custom\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525187 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgf7k\" (UniqueName: \"kubernetes.io/projected/f3cb9fa5-3711-4f0a-830b-e2d210caada6-kube-api-access-xgf7k\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525206 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-config-data-custom\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525339 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-config-data\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525442 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-combined-ca-bundle\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525500 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-combined-ca-bundle\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525798 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlwhb\" (UniqueName: \"kubernetes.io/projected/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-kube-api-access-wlwhb\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525843 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-config-data-custom\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.525875 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-config-data\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.534977 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-combined-ca-bundle\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.535464 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-config-data\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.536674 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-config-data-custom\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.538605 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-combined-ca-bundle\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.544714 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-config-data\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.546716 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlwhb\" (UniqueName: \"kubernetes.io/projected/5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17-kube-api-access-wlwhb\") pod \"heat-api-54d757dd4f-264jb\" (UID: \"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17\") " pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.547453 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3cb9fa5-3711-4f0a-830b-e2d210caada6-config-data-custom\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.550429 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgf7k\" (UniqueName: \"kubernetes.io/projected/f3cb9fa5-3711-4f0a-830b-e2d210caada6-kube-api-access-xgf7k\") pod \"heat-engine-59d5d97855-8b85w\" (UID: \"f3cb9fa5-3711-4f0a-830b-e2d210caada6\") " pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.567417 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.629527 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47kgv\" (UniqueName: \"kubernetes.io/projected/dfeb90b7-9abc-4f02-8a73-2c6df751c902-kube-api-access-47kgv\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.630145 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-combined-ca-bundle\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.630220 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-config-data-custom\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.630300 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-config-data\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.641075 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-config-data\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.643112 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-config-data-custom\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.643559 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.655360 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47kgv\" (UniqueName: \"kubernetes.io/projected/dfeb90b7-9abc-4f02-8a73-2c6df751c902-kube-api-access-47kgv\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.657578 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfeb90b7-9abc-4f02-8a73-2c6df751c902-combined-ca-bundle\") pod \"heat-cfnapi-59984465f-5qgvh\" (UID: \"dfeb90b7-9abc-4f02-8a73-2c6df751c902\") " pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.714832 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:36 crc kubenswrapper[4935]: I1005 08:58:36.715627 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-75bb459497-g9b7r" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.106140 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-59d5d97855-8b85w"] Oct 05 08:58:37 crc kubenswrapper[4935]: W1005 08:58:37.114997 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3cb9fa5_3711_4f0a_830b_e2d210caada6.slice/crio-301000c37025f6f051b65b68bc0386f032b8c336ec6d15e24f343c915d33d5c1 WatchSource:0}: Error finding container 301000c37025f6f051b65b68bc0386f032b8c336ec6d15e24f343c915d33d5c1: Status 404 returned error can't find the container with id 301000c37025f6f051b65b68bc0386f032b8c336ec6d15e24f343c915d33d5c1 Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.262672 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-54d757dd4f-264jb"] Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.347877 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-59984465f-5qgvh"] Oct 05 08:58:37 crc kubenswrapper[4935]: W1005 08:58:37.354850 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfeb90b7_9abc_4f02_8a73_2c6df751c902.slice/crio-902ec5c4e3647fa0ec221c432ac5546d49512e5f5486db80cf8dda8cdb9c3870 WatchSource:0}: Error finding container 902ec5c4e3647fa0ec221c432ac5546d49512e5f5486db80cf8dda8cdb9c3870: Status 404 returned error can't find the container with id 902ec5c4e3647fa0ec221c432ac5546d49512e5f5486db80cf8dda8cdb9c3870 Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.801575 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-54d757dd4f-264jb" event={"ID":"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17","Type":"ContainerStarted","Data":"98e9a83f0d83f2c13a86e57dc52787865a558f93f60a087a78091c9a208d0b94"} Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.804029 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-59d5d97855-8b85w" event={"ID":"f3cb9fa5-3711-4f0a-830b-e2d210caada6","Type":"ContainerStarted","Data":"314be18e2c07512cdf1a1e952a3b71502ecb7569b51071401655699a0045ab0e"} Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.804073 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-59d5d97855-8b85w" event={"ID":"f3cb9fa5-3711-4f0a-830b-e2d210caada6","Type":"ContainerStarted","Data":"301000c37025f6f051b65b68bc0386f032b8c336ec6d15e24f343c915d33d5c1"} Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.804114 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.805761 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59984465f-5qgvh" event={"ID":"dfeb90b7-9abc-4f02-8a73-2c6df751c902","Type":"ContainerStarted","Data":"902ec5c4e3647fa0ec221c432ac5546d49512e5f5486db80cf8dda8cdb9c3870"} Oct 05 08:58:37 crc kubenswrapper[4935]: I1005 08:58:37.823441 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-59d5d97855-8b85w" podStartSLOduration=1.8234231319999998 podStartE2EDuration="1.823423132s" podCreationTimestamp="2025-10-05 08:58:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:58:37.820029492 +0000 UTC m=+7551.702655942" watchObservedRunningTime="2025-10-05 08:58:37.823423132 +0000 UTC m=+7551.706049592" Oct 05 08:58:39 crc kubenswrapper[4935]: I1005 08:58:39.842110 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-54d757dd4f-264jb" event={"ID":"5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17","Type":"ContainerStarted","Data":"b49a58792bad7eb624bd8ba74665929a7ff773ce8484d6add15c566367236e6d"} Oct 05 08:58:39 crc kubenswrapper[4935]: I1005 08:58:39.843718 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:39 crc kubenswrapper[4935]: I1005 08:58:39.845290 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59984465f-5qgvh" event={"ID":"dfeb90b7-9abc-4f02-8a73-2c6df751c902","Type":"ContainerStarted","Data":"f0a60952463cb8280752374064985d321f0cf25d2a77364fb3c97e472bd98742"} Oct 05 08:58:39 crc kubenswrapper[4935]: I1005 08:58:39.845787 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:39 crc kubenswrapper[4935]: I1005 08:58:39.862622 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-54d757dd4f-264jb" podStartSLOduration=2.384354155 podStartE2EDuration="3.862607556s" podCreationTimestamp="2025-10-05 08:58:36 +0000 UTC" firstStartedPulling="2025-10-05 08:58:37.277068467 +0000 UTC m=+7551.159694927" lastFinishedPulling="2025-10-05 08:58:38.755321858 +0000 UTC m=+7552.637948328" observedRunningTime="2025-10-05 08:58:39.860982643 +0000 UTC m=+7553.743609103" watchObservedRunningTime="2025-10-05 08:58:39.862607556 +0000 UTC m=+7553.745234006" Oct 05 08:58:39 crc kubenswrapper[4935]: I1005 08:58:39.887036 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-59984465f-5qgvh" podStartSLOduration=2.464327265 podStartE2EDuration="3.887016363s" podCreationTimestamp="2025-10-05 08:58:36 +0000 UTC" firstStartedPulling="2025-10-05 08:58:37.356582555 +0000 UTC m=+7551.239209015" lastFinishedPulling="2025-10-05 08:58:38.779271653 +0000 UTC m=+7552.661898113" observedRunningTime="2025-10-05 08:58:39.879841263 +0000 UTC m=+7553.762467743" watchObservedRunningTime="2025-10-05 08:58:39.887016363 +0000 UTC m=+7553.769642823" Oct 05 08:58:41 crc kubenswrapper[4935]: I1005 08:58:41.041548 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d00d-account-create-vc9qx"] Oct 05 08:58:41 crc kubenswrapper[4935]: I1005 08:58:41.055386 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d00d-account-create-vc9qx"] Oct 05 08:58:42 crc kubenswrapper[4935]: I1005 08:58:42.792222 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec8da432-4737-4b96-97bb-7e163a809f75" path="/var/lib/kubelet/pods/ec8da432-4737-4b96-97bb-7e163a809f75/volumes" Oct 05 08:58:46 crc kubenswrapper[4935]: I1005 08:58:46.713858 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-75bb459497-g9b7r" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 05 08:58:47 crc kubenswrapper[4935]: I1005 08:58:47.951517 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-54d757dd4f-264jb" Oct 05 08:58:47 crc kubenswrapper[4935]: I1005 08:58:47.986293 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-59984465f-5qgvh" Oct 05 08:58:50 crc kubenswrapper[4935]: I1005 08:58:50.029517 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-d7twt"] Oct 05 08:58:50 crc kubenswrapper[4935]: I1005 08:58:50.039693 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-d7twt"] Oct 05 08:58:50 crc kubenswrapper[4935]: I1005 08:58:50.795846 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a03a073a-7848-4ba1-b73a-ed3afa62a1e8" path="/var/lib/kubelet/pods/a03a073a-7848-4ba1-b73a-ed3afa62a1e8/volumes" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.746875 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gx4n7"] Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.756051 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.766506 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gx4n7"] Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.869231 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-utilities\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.869321 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-catalog-content\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.869382 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vz62\" (UniqueName: \"kubernetes.io/projected/e5106f7d-6147-4d4f-936b-036c9fa4e20f-kube-api-access-5vz62\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.971044 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-catalog-content\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.971124 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vz62\" (UniqueName: \"kubernetes.io/projected/e5106f7d-6147-4d4f-936b-036c9fa4e20f-kube-api-access-5vz62\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.971240 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-utilities\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.971585 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-catalog-content\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.971612 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-utilities\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:51 crc kubenswrapper[4935]: I1005 08:58:51.994276 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vz62\" (UniqueName: \"kubernetes.io/projected/e5106f7d-6147-4d4f-936b-036c9fa4e20f-kube-api-access-5vz62\") pod \"redhat-marketplace-gx4n7\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:52 crc kubenswrapper[4935]: I1005 08:58:52.076327 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:58:52 crc kubenswrapper[4935]: I1005 08:58:52.663101 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gx4n7"] Oct 05 08:58:53 crc kubenswrapper[4935]: I1005 08:58:53.009555 4935 generic.go:334] "Generic (PLEG): container finished" podID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerID="5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165" exitCode=0 Oct 05 08:58:53 crc kubenswrapper[4935]: I1005 08:58:53.009600 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx4n7" event={"ID":"e5106f7d-6147-4d4f-936b-036c9fa4e20f","Type":"ContainerDied","Data":"5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165"} Oct 05 08:58:53 crc kubenswrapper[4935]: I1005 08:58:53.009624 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx4n7" event={"ID":"e5106f7d-6147-4d4f-936b-036c9fa4e20f","Type":"ContainerStarted","Data":"3e8e6c9d442a4c8af587b08408ceb7a7159a20074aaf1dc6f113276f95cbf4f2"} Oct 05 08:58:54 crc kubenswrapper[4935]: I1005 08:58:54.023046 4935 generic.go:334] "Generic (PLEG): container finished" podID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerID="6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f" exitCode=0 Oct 05 08:58:54 crc kubenswrapper[4935]: I1005 08:58:54.023125 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx4n7" event={"ID":"e5106f7d-6147-4d4f-936b-036c9fa4e20f","Type":"ContainerDied","Data":"6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f"} Oct 05 08:58:55 crc kubenswrapper[4935]: I1005 08:58:55.043764 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx4n7" event={"ID":"e5106f7d-6147-4d4f-936b-036c9fa4e20f","Type":"ContainerStarted","Data":"bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce"} Oct 05 08:58:55 crc kubenswrapper[4935]: I1005 08:58:55.076064 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gx4n7" podStartSLOduration=2.688729801 podStartE2EDuration="4.076034053s" podCreationTimestamp="2025-10-05 08:58:51 +0000 UTC" firstStartedPulling="2025-10-05 08:58:53.011298331 +0000 UTC m=+7566.893924791" lastFinishedPulling="2025-10-05 08:58:54.398602543 +0000 UTC m=+7568.281229043" observedRunningTime="2025-10-05 08:58:55.065277218 +0000 UTC m=+7568.947903678" watchObservedRunningTime="2025-10-05 08:58:55.076034053 +0000 UTC m=+7568.958660513" Oct 05 08:58:56 crc kubenswrapper[4935]: I1005 08:58:56.594732 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-59d5d97855-8b85w" Oct 05 08:58:56 crc kubenswrapper[4935]: I1005 08:58:56.713984 4935 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-75bb459497-g9b7r" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 05 08:58:56 crc kubenswrapper[4935]: I1005 08:58:56.714085 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:58:58 crc kubenswrapper[4935]: I1005 08:58:58.973785 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.081015 4935 generic.go:334] "Generic (PLEG): container finished" podID="2cdeadef-85fc-4477-8352-dc671812251e" containerID="cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38" exitCode=137 Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.081061 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bb459497-g9b7r" event={"ID":"2cdeadef-85fc-4477-8352-dc671812251e","Type":"ContainerDied","Data":"cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38"} Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.081067 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75bb459497-g9b7r" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.081098 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bb459497-g9b7r" event={"ID":"2cdeadef-85fc-4477-8352-dc671812251e","Type":"ContainerDied","Data":"4a55228b78f7604e85790ff6894ab2d90ad03b932d74972fecab174d8200c6d1"} Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.081119 4935 scope.go:117] "RemoveContainer" containerID="4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.095015 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm4np\" (UniqueName: \"kubernetes.io/projected/2cdeadef-85fc-4477-8352-dc671812251e-kube-api-access-lm4np\") pod \"2cdeadef-85fc-4477-8352-dc671812251e\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.096252 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-scripts\") pod \"2cdeadef-85fc-4477-8352-dc671812251e\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.096288 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cdeadef-85fc-4477-8352-dc671812251e-logs\") pod \"2cdeadef-85fc-4477-8352-dc671812251e\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.096349 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cdeadef-85fc-4477-8352-dc671812251e-horizon-secret-key\") pod \"2cdeadef-85fc-4477-8352-dc671812251e\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.096468 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-config-data\") pod \"2cdeadef-85fc-4477-8352-dc671812251e\" (UID: \"2cdeadef-85fc-4477-8352-dc671812251e\") " Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.097574 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cdeadef-85fc-4477-8352-dc671812251e-logs" (OuterVolumeSpecName: "logs") pod "2cdeadef-85fc-4477-8352-dc671812251e" (UID: "2cdeadef-85fc-4477-8352-dc671812251e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.101226 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cdeadef-85fc-4477-8352-dc671812251e-kube-api-access-lm4np" (OuterVolumeSpecName: "kube-api-access-lm4np") pod "2cdeadef-85fc-4477-8352-dc671812251e" (UID: "2cdeadef-85fc-4477-8352-dc671812251e"). InnerVolumeSpecName "kube-api-access-lm4np". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.101598 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cdeadef-85fc-4477-8352-dc671812251e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2cdeadef-85fc-4477-8352-dc671812251e" (UID: "2cdeadef-85fc-4477-8352-dc671812251e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.130324 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-scripts" (OuterVolumeSpecName: "scripts") pod "2cdeadef-85fc-4477-8352-dc671812251e" (UID: "2cdeadef-85fc-4477-8352-dc671812251e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.132495 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-config-data" (OuterVolumeSpecName: "config-data") pod "2cdeadef-85fc-4477-8352-dc671812251e" (UID: "2cdeadef-85fc-4477-8352-dc671812251e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.198813 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.198843 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm4np\" (UniqueName: \"kubernetes.io/projected/2cdeadef-85fc-4477-8352-dc671812251e-kube-api-access-lm4np\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.198854 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cdeadef-85fc-4477-8352-dc671812251e-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.198863 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cdeadef-85fc-4477-8352-dc671812251e-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.198873 4935 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cdeadef-85fc-4477-8352-dc671812251e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.244162 4935 scope.go:117] "RemoveContainer" containerID="cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.278008 4935 scope.go:117] "RemoveContainer" containerID="4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73" Oct 05 08:58:59 crc kubenswrapper[4935]: E1005 08:58:59.278522 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73\": container with ID starting with 4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73 not found: ID does not exist" containerID="4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.278570 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73"} err="failed to get container status \"4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73\": rpc error: code = NotFound desc = could not find container \"4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73\": container with ID starting with 4993fea515950e631450f33d30b55790759991b1bec2982d21b2df6e996dfa73 not found: ID does not exist" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.278597 4935 scope.go:117] "RemoveContainer" containerID="cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38" Oct 05 08:58:59 crc kubenswrapper[4935]: E1005 08:58:59.279095 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38\": container with ID starting with cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38 not found: ID does not exist" containerID="cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.279135 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38"} err="failed to get container status \"cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38\": rpc error: code = NotFound desc = could not find container \"cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38\": container with ID starting with cc7677d3a9e538280bb9bd79496b9ceeee1bce2592ce479c532a5c8047c40e38 not found: ID does not exist" Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.410068 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75bb459497-g9b7r"] Oct 05 08:58:59 crc kubenswrapper[4935]: I1005 08:58:59.419432 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-75bb459497-g9b7r"] Oct 05 08:59:00 crc kubenswrapper[4935]: I1005 08:59:00.550565 4935 scope.go:117] "RemoveContainer" containerID="135e09dd83ca1be0f4c18e6764f986b4108307c6c20807a70682b635819d1bcd" Oct 05 08:59:00 crc kubenswrapper[4935]: I1005 08:59:00.594569 4935 scope.go:117] "RemoveContainer" containerID="5eadd1cd946c5586d843a75abad297459c95b85f169b14714f69c2dc39e19b72" Oct 05 08:59:00 crc kubenswrapper[4935]: I1005 08:59:00.656045 4935 scope.go:117] "RemoveContainer" containerID="cce76dcc3fa400c3a3b0819021f5ac69981210a476f595caa7c2dc2fe7e616b5" Oct 05 08:59:00 crc kubenswrapper[4935]: I1005 08:59:00.676463 4935 scope.go:117] "RemoveContainer" containerID="afc6a31d527b8f997602ffb199a324e35da96ba522ed08d42332ed13903bd713" Oct 05 08:59:00 crc kubenswrapper[4935]: I1005 08:59:00.799589 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cdeadef-85fc-4477-8352-dc671812251e" path="/var/lib/kubelet/pods/2cdeadef-85fc-4477-8352-dc671812251e/volumes" Oct 05 08:59:02 crc kubenswrapper[4935]: I1005 08:59:02.077111 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:59:02 crc kubenswrapper[4935]: I1005 08:59:02.077525 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:59:02 crc kubenswrapper[4935]: I1005 08:59:02.155047 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:59:02 crc kubenswrapper[4935]: I1005 08:59:02.239490 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:59:02 crc kubenswrapper[4935]: I1005 08:59:02.396463 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gx4n7"] Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.130056 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gx4n7" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerName="registry-server" containerID="cri-o://bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce" gracePeriod=2 Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.620720 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.810277 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-utilities\") pod \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.810728 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-catalog-content\") pod \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.810862 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vz62\" (UniqueName: \"kubernetes.io/projected/e5106f7d-6147-4d4f-936b-036c9fa4e20f-kube-api-access-5vz62\") pod \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\" (UID: \"e5106f7d-6147-4d4f-936b-036c9fa4e20f\") " Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.811057 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-utilities" (OuterVolumeSpecName: "utilities") pod "e5106f7d-6147-4d4f-936b-036c9fa4e20f" (UID: "e5106f7d-6147-4d4f-936b-036c9fa4e20f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.811718 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.819910 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5106f7d-6147-4d4f-936b-036c9fa4e20f-kube-api-access-5vz62" (OuterVolumeSpecName: "kube-api-access-5vz62") pod "e5106f7d-6147-4d4f-936b-036c9fa4e20f" (UID: "e5106f7d-6147-4d4f-936b-036c9fa4e20f"). InnerVolumeSpecName "kube-api-access-5vz62". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.824998 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5106f7d-6147-4d4f-936b-036c9fa4e20f" (UID: "e5106f7d-6147-4d4f-936b-036c9fa4e20f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.913483 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5106f7d-6147-4d4f-936b-036c9fa4e20f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:04 crc kubenswrapper[4935]: I1005 08:59:04.915147 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vz62\" (UniqueName: \"kubernetes.io/projected/e5106f7d-6147-4d4f-936b-036c9fa4e20f-kube-api-access-5vz62\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.141829 4935 generic.go:334] "Generic (PLEG): container finished" podID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerID="bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce" exitCode=0 Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.141921 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx4n7" event={"ID":"e5106f7d-6147-4d4f-936b-036c9fa4e20f","Type":"ContainerDied","Data":"bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce"} Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.141941 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gx4n7" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.142001 4935 scope.go:117] "RemoveContainer" containerID="bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.141983 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gx4n7" event={"ID":"e5106f7d-6147-4d4f-936b-036c9fa4e20f","Type":"ContainerDied","Data":"3e8e6c9d442a4c8af587b08408ceb7a7159a20074aaf1dc6f113276f95cbf4f2"} Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.166248 4935 scope.go:117] "RemoveContainer" containerID="6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.181221 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gx4n7"] Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.209335 4935 scope.go:117] "RemoveContainer" containerID="5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.222627 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gx4n7"] Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.255496 4935 scope.go:117] "RemoveContainer" containerID="bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce" Oct 05 08:59:05 crc kubenswrapper[4935]: E1005 08:59:05.256085 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce\": container with ID starting with bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce not found: ID does not exist" containerID="bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.256128 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce"} err="failed to get container status \"bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce\": rpc error: code = NotFound desc = could not find container \"bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce\": container with ID starting with bc52e79a002609f8fce83cab307e97cff083d101a2bd8f35ccb759a8550d94ce not found: ID does not exist" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.256150 4935 scope.go:117] "RemoveContainer" containerID="6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f" Oct 05 08:59:05 crc kubenswrapper[4935]: E1005 08:59:05.258535 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f\": container with ID starting with 6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f not found: ID does not exist" containerID="6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.258571 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f"} err="failed to get container status \"6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f\": rpc error: code = NotFound desc = could not find container \"6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f\": container with ID starting with 6604593f88a19d760772a209fa7e71b8c581b121c4848c911fa4631ef6fa663f not found: ID does not exist" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.258589 4935 scope.go:117] "RemoveContainer" containerID="5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165" Oct 05 08:59:05 crc kubenswrapper[4935]: E1005 08:59:05.259025 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165\": container with ID starting with 5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165 not found: ID does not exist" containerID="5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165" Oct 05 08:59:05 crc kubenswrapper[4935]: I1005 08:59:05.259055 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165"} err="failed to get container status \"5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165\": rpc error: code = NotFound desc = could not find container \"5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165\": container with ID starting with 5aaba2287eac3fd6614acb32492471d921bf99d08c54316e4e504b5b5938c165 not found: ID does not exist" Oct 05 08:59:06 crc kubenswrapper[4935]: I1005 08:59:06.788462 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" path="/var/lib/kubelet/pods/e5106f7d-6147-4d4f-936b-036c9fa4e20f/volumes" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.257845 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg"] Oct 05 08:59:10 crc kubenswrapper[4935]: E1005 08:59:10.259581 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerName="registry-server" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.259616 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerName="registry-server" Oct 05 08:59:10 crc kubenswrapper[4935]: E1005 08:59:10.259637 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerName="extract-utilities" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.259646 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerName="extract-utilities" Oct 05 08:59:10 crc kubenswrapper[4935]: E1005 08:59:10.259660 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerName="extract-content" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.259668 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerName="extract-content" Oct 05 08:59:10 crc kubenswrapper[4935]: E1005 08:59:10.259700 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon-log" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.259709 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon-log" Oct 05 08:59:10 crc kubenswrapper[4935]: E1005 08:59:10.259727 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.259735 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.259993 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.260023 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cdeadef-85fc-4477-8352-dc671812251e" containerName="horizon-log" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.260045 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5106f7d-6147-4d4f-936b-036c9fa4e20f" containerName="registry-server" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.261739 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.263617 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.276572 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg"] Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.322744 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.322809 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.322862 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87ghs\" (UniqueName: \"kubernetes.io/projected/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-kube-api-access-87ghs\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.424492 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.424558 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.424616 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87ghs\" (UniqueName: \"kubernetes.io/projected/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-kube-api-access-87ghs\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.425388 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.425868 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.443444 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87ghs\" (UniqueName: \"kubernetes.io/projected/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-kube-api-access-87ghs\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:10 crc kubenswrapper[4935]: I1005 08:59:10.588802 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:11 crc kubenswrapper[4935]: I1005 08:59:11.064685 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg"] Oct 05 08:59:11 crc kubenswrapper[4935]: I1005 08:59:11.199950 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" event={"ID":"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f","Type":"ContainerStarted","Data":"1c80110da747bbf61a748c170752329714fe8d5b412fda76680b07bdc4af33ca"} Oct 05 08:59:12 crc kubenswrapper[4935]: I1005 08:59:12.213654 4935 generic.go:334] "Generic (PLEG): container finished" podID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerID="9f681f7a8d0cdac8adcec0bd2b35274c3da9ab3670e2f3cabbb99a77fd1bc377" exitCode=0 Oct 05 08:59:12 crc kubenswrapper[4935]: I1005 08:59:12.213715 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" event={"ID":"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f","Type":"ContainerDied","Data":"9f681f7a8d0cdac8adcec0bd2b35274c3da9ab3670e2f3cabbb99a77fd1bc377"} Oct 05 08:59:14 crc kubenswrapper[4935]: I1005 08:59:14.239632 4935 generic.go:334] "Generic (PLEG): container finished" podID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerID="ec3a9e75f5d5ac22bf90d9abca6c4ec6b6c6f5acf647379bd6d1c69a526d17ad" exitCode=0 Oct 05 08:59:14 crc kubenswrapper[4935]: I1005 08:59:14.239773 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" event={"ID":"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f","Type":"ContainerDied","Data":"ec3a9e75f5d5ac22bf90d9abca6c4ec6b6c6f5acf647379bd6d1c69a526d17ad"} Oct 05 08:59:15 crc kubenswrapper[4935]: I1005 08:59:15.261621 4935 generic.go:334] "Generic (PLEG): container finished" podID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerID="d433d798b320c8fbc475be4ce3bafc3ecfe24aef8d707940c98a4fc0b645edda" exitCode=0 Oct 05 08:59:15 crc kubenswrapper[4935]: I1005 08:59:15.261766 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" event={"ID":"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f","Type":"ContainerDied","Data":"d433d798b320c8fbc475be4ce3bafc3ecfe24aef8d707940c98a4fc0b645edda"} Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.666363 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.782299 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-util\") pod \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.782384 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-bundle\") pod \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.782431 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87ghs\" (UniqueName: \"kubernetes.io/projected/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-kube-api-access-87ghs\") pod \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\" (UID: \"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f\") " Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.784449 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-bundle" (OuterVolumeSpecName: "bundle") pod "4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" (UID: "4f0ed86d-4848-4b38-b2d0-39a9dceaca1f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.788251 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-kube-api-access-87ghs" (OuterVolumeSpecName: "kube-api-access-87ghs") pod "4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" (UID: "4f0ed86d-4848-4b38-b2d0-39a9dceaca1f"). InnerVolumeSpecName "kube-api-access-87ghs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.796118 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-util" (OuterVolumeSpecName: "util") pod "4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" (UID: "4f0ed86d-4848-4b38-b2d0-39a9dceaca1f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.884208 4935 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-util\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.884249 4935 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:16 crc kubenswrapper[4935]: I1005 08:59:16.884259 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87ghs\" (UniqueName: \"kubernetes.io/projected/4f0ed86d-4848-4b38-b2d0-39a9dceaca1f-kube-api-access-87ghs\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:17 crc kubenswrapper[4935]: I1005 08:59:17.284100 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" event={"ID":"4f0ed86d-4848-4b38-b2d0-39a9dceaca1f","Type":"ContainerDied","Data":"1c80110da747bbf61a748c170752329714fe8d5b412fda76680b07bdc4af33ca"} Oct 05 08:59:17 crc kubenswrapper[4935]: I1005 08:59:17.284137 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c80110da747bbf61a748c170752329714fe8d5b412fda76680b07bdc4af33ca" Oct 05 08:59:17 crc kubenswrapper[4935]: I1005 08:59:17.284164 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.927544 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft"] Oct 05 08:59:27 crc kubenswrapper[4935]: E1005 08:59:27.928420 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerName="util" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.928433 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerName="util" Oct 05 08:59:27 crc kubenswrapper[4935]: E1005 08:59:27.928443 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerName="pull" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.928449 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerName="pull" Oct 05 08:59:27 crc kubenswrapper[4935]: E1005 08:59:27.928486 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerName="extract" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.928494 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerName="extract" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.928658 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0ed86d-4848-4b38-b2d0-39a9dceaca1f" containerName="extract" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.929327 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.935382 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.935487 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-7lwhz" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.935542 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.950146 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft"] Oct 05 08:59:27 crc kubenswrapper[4935]: I1005 08:59:27.978095 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f9nv\" (UniqueName: \"kubernetes.io/projected/d0952f68-f026-4b91-bf20-41c0f6b3ae09-kube-api-access-7f9nv\") pod \"obo-prometheus-operator-7c8cf85677-t8cft\" (UID: \"d0952f68-f026-4b91-bf20-41c0f6b3ae09\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.054845 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt"] Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.056016 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.058072 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.060724 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-bkg78" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.066473 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt"] Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.078286 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm"] Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.079400 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f9nv\" (UniqueName: \"kubernetes.io/projected/d0952f68-f026-4b91-bf20-41c0f6b3ae09-kube-api-access-7f9nv\") pod \"obo-prometheus-operator-7c8cf85677-t8cft\" (UID: \"d0952f68-f026-4b91-bf20-41c0f6b3ae09\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.080298 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.098745 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm"] Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.117647 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f9nv\" (UniqueName: \"kubernetes.io/projected/d0952f68-f026-4b91-bf20-41c0f6b3ae09-kube-api-access-7f9nv\") pod \"obo-prometheus-operator-7c8cf85677-t8cft\" (UID: \"d0952f68-f026-4b91-bf20-41c0f6b3ae09\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.183319 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56891111-0316-4485-9beb-78e3d36385be-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-bckjm\" (UID: \"56891111-0316-4485-9beb-78e3d36385be\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.183664 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bed510a3-6d84-402d-9c73-22adbae5329f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-rxngt\" (UID: \"bed510a3-6d84-402d-9c73-22adbae5329f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.183727 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bed510a3-6d84-402d-9c73-22adbae5329f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-rxngt\" (UID: \"bed510a3-6d84-402d-9c73-22adbae5329f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.183806 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56891111-0316-4485-9beb-78e3d36385be-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-bckjm\" (UID: \"56891111-0316-4485-9beb-78e3d36385be\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.222067 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qhnds"] Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.223385 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.232251 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-cmqb6" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.232480 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.248890 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qhnds"] Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.254535 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.304085 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgkqv\" (UniqueName: \"kubernetes.io/projected/4018d5a3-249d-4dcb-a528-504a9ce098bf-kube-api-access-sgkqv\") pod \"observability-operator-cc5f78dfc-qhnds\" (UID: \"4018d5a3-249d-4dcb-a528-504a9ce098bf\") " pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.304204 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56891111-0316-4485-9beb-78e3d36385be-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-bckjm\" (UID: \"56891111-0316-4485-9beb-78e3d36385be\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.304238 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4018d5a3-249d-4dcb-a528-504a9ce098bf-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qhnds\" (UID: \"4018d5a3-249d-4dcb-a528-504a9ce098bf\") " pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.304269 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bed510a3-6d84-402d-9c73-22adbae5329f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-rxngt\" (UID: \"bed510a3-6d84-402d-9c73-22adbae5329f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.304392 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bed510a3-6d84-402d-9c73-22adbae5329f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-rxngt\" (UID: \"bed510a3-6d84-402d-9c73-22adbae5329f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.304591 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56891111-0316-4485-9beb-78e3d36385be-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-bckjm\" (UID: \"56891111-0316-4485-9beb-78e3d36385be\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.309414 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56891111-0316-4485-9beb-78e3d36385be-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-bckjm\" (UID: \"56891111-0316-4485-9beb-78e3d36385be\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.309423 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56891111-0316-4485-9beb-78e3d36385be-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-bckjm\" (UID: \"56891111-0316-4485-9beb-78e3d36385be\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.311365 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bed510a3-6d84-402d-9c73-22adbae5329f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-rxngt\" (UID: \"bed510a3-6d84-402d-9c73-22adbae5329f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.311700 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bed510a3-6d84-402d-9c73-22adbae5329f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-649566fff-rxngt\" (UID: \"bed510a3-6d84-402d-9c73-22adbae5329f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.375664 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ztdwx"] Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.383142 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.386047 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.387969 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-4xhwj" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.397522 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ztdwx"] Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.412003 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4018d5a3-249d-4dcb-a528-504a9ce098bf-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qhnds\" (UID: \"4018d5a3-249d-4dcb-a528-504a9ce098bf\") " pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.412145 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgkqv\" (UniqueName: \"kubernetes.io/projected/4018d5a3-249d-4dcb-a528-504a9ce098bf-kube-api-access-sgkqv\") pod \"observability-operator-cc5f78dfc-qhnds\" (UID: \"4018d5a3-249d-4dcb-a528-504a9ce098bf\") " pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.413109 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.430677 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4018d5a3-249d-4dcb-a528-504a9ce098bf-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qhnds\" (UID: \"4018d5a3-249d-4dcb-a528-504a9ce098bf\") " pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.444572 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgkqv\" (UniqueName: \"kubernetes.io/projected/4018d5a3-249d-4dcb-a528-504a9ce098bf-kube-api-access-sgkqv\") pod \"observability-operator-cc5f78dfc-qhnds\" (UID: \"4018d5a3-249d-4dcb-a528-504a9ce098bf\") " pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.517005 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/edf2f10d-a59a-4d08-81b2-b86892f02e54-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ztdwx\" (UID: \"edf2f10d-a59a-4d08-81b2-b86892f02e54\") " pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.517087 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mj7j\" (UniqueName: \"kubernetes.io/projected/edf2f10d-a59a-4d08-81b2-b86892f02e54-kube-api-access-6mj7j\") pod \"perses-operator-54bc95c9fb-ztdwx\" (UID: \"edf2f10d-a59a-4d08-81b2-b86892f02e54\") " pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.556963 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.619416 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/edf2f10d-a59a-4d08-81b2-b86892f02e54-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ztdwx\" (UID: \"edf2f10d-a59a-4d08-81b2-b86892f02e54\") " pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.619473 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mj7j\" (UniqueName: \"kubernetes.io/projected/edf2f10d-a59a-4d08-81b2-b86892f02e54-kube-api-access-6mj7j\") pod \"perses-operator-54bc95c9fb-ztdwx\" (UID: \"edf2f10d-a59a-4d08-81b2-b86892f02e54\") " pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.620517 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/edf2f10d-a59a-4d08-81b2-b86892f02e54-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ztdwx\" (UID: \"edf2f10d-a59a-4d08-81b2-b86892f02e54\") " pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.649922 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mj7j\" (UniqueName: \"kubernetes.io/projected/edf2f10d-a59a-4d08-81b2-b86892f02e54-kube-api-access-6mj7j\") pod \"perses-operator-54bc95c9fb-ztdwx\" (UID: \"edf2f10d-a59a-4d08-81b2-b86892f02e54\") " pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:28 crc kubenswrapper[4935]: I1005 08:59:28.792098 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.050956 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft"] Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.142420 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt"] Oct 05 08:59:29 crc kubenswrapper[4935]: W1005 08:59:29.143020 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbed510a3_6d84_402d_9c73_22adbae5329f.slice/crio-11f6b3da4e7bf4dff01f19f55af598a9c70448df057734e8d300972cbca1fc8a WatchSource:0}: Error finding container 11f6b3da4e7bf4dff01f19f55af598a9c70448df057734e8d300972cbca1fc8a: Status 404 returned error can't find the container with id 11f6b3da4e7bf4dff01f19f55af598a9c70448df057734e8d300972cbca1fc8a Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.370071 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm"] Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.384209 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ztdwx"] Oct 05 08:59:29 crc kubenswrapper[4935]: W1005 08:59:29.404351 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedf2f10d_a59a_4d08_81b2_b86892f02e54.slice/crio-9b58da67a67db87f9ce59150aeb1dc2461ff8c6c3039f173f6e5cd8ff0017197 WatchSource:0}: Error finding container 9b58da67a67db87f9ce59150aeb1dc2461ff8c6c3039f173f6e5cd8ff0017197: Status 404 returned error can't find the container with id 9b58da67a67db87f9ce59150aeb1dc2461ff8c6c3039f173f6e5cd8ff0017197 Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.426938 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft" event={"ID":"d0952f68-f026-4b91-bf20-41c0f6b3ae09","Type":"ContainerStarted","Data":"c4276f44cebd0898f6fcf91aa7a31889e383f62fea0924c060271ce5c2dc6744"} Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.430048 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" event={"ID":"bed510a3-6d84-402d-9c73-22adbae5329f","Type":"ContainerStarted","Data":"11f6b3da4e7bf4dff01f19f55af598a9c70448df057734e8d300972cbca1fc8a"} Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.431568 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" event={"ID":"edf2f10d-a59a-4d08-81b2-b86892f02e54","Type":"ContainerStarted","Data":"9b58da67a67db87f9ce59150aeb1dc2461ff8c6c3039f173f6e5cd8ff0017197"} Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.432568 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" event={"ID":"56891111-0316-4485-9beb-78e3d36385be","Type":"ContainerStarted","Data":"29b033a4aec4f5c0808cd6326bd694ca0e5ccc90af2964da6f4bceee9873cc0a"} Oct 05 08:59:29 crc kubenswrapper[4935]: I1005 08:59:29.556760 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qhnds"] Oct 05 08:59:30 crc kubenswrapper[4935]: I1005 08:59:30.446420 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" event={"ID":"4018d5a3-249d-4dcb-a528-504a9ce098bf","Type":"ContainerStarted","Data":"3408085b5796cc9ea029315a066be41c8e81004f9e64cd2836758cada597194f"} Oct 05 08:59:32 crc kubenswrapper[4935]: I1005 08:59:32.036543 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-f849r"] Oct 05 08:59:32 crc kubenswrapper[4935]: I1005 08:59:32.047474 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-f849r"] Oct 05 08:59:32 crc kubenswrapper[4935]: I1005 08:59:32.787112 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b87e8ba5-cc47-41b7-b1dd-155a2942ce6a" path="/var/lib/kubelet/pods/b87e8ba5-cc47-41b7-b1dd-155a2942ce6a/volumes" Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.504276 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft" event={"ID":"d0952f68-f026-4b91-bf20-41c0f6b3ae09","Type":"ContainerStarted","Data":"05b4f831c15e11089a8fb764d09c941ae40b8e47b471277ecba47a6e1f200820"} Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.505998 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" event={"ID":"bed510a3-6d84-402d-9c73-22adbae5329f","Type":"ContainerStarted","Data":"a644c543570cec29fe1213dca761bdd1a9d7cf411c4186cc036c576a50047ac6"} Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.507734 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" event={"ID":"edf2f10d-a59a-4d08-81b2-b86892f02e54","Type":"ContainerStarted","Data":"c85b7362de115002eda46e9e34dce36531f885b090dc6625bc8d7d1f513689be"} Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.507808 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.509773 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" event={"ID":"56891111-0316-4485-9beb-78e3d36385be","Type":"ContainerStarted","Data":"6352c8fe8848f246b1fddafed0ab081d6c58ea93541a9a67a163eaab20733cdc"} Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.511253 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" event={"ID":"4018d5a3-249d-4dcb-a528-504a9ce098bf","Type":"ContainerStarted","Data":"f1f0cab3cc305c697b33fb9ba03365d28d0bed6d00bc7a6f471af84c4301179d"} Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.511466 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.551558 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-t8cft" podStartSLOduration=3.097970065 podStartE2EDuration="10.551537249s" podCreationTimestamp="2025-10-05 08:59:27 +0000 UTC" firstStartedPulling="2025-10-05 08:59:29.062001779 +0000 UTC m=+7602.944628239" lastFinishedPulling="2025-10-05 08:59:36.515568963 +0000 UTC m=+7610.398195423" observedRunningTime="2025-10-05 08:59:37.526185037 +0000 UTC m=+7611.408811497" watchObservedRunningTime="2025-10-05 08:59:37.551537249 +0000 UTC m=+7611.434163709" Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.552592 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-rxngt" podStartSLOduration=2.182766323 podStartE2EDuration="9.552586977s" podCreationTimestamp="2025-10-05 08:59:28 +0000 UTC" firstStartedPulling="2025-10-05 08:59:29.145753679 +0000 UTC m=+7603.028380129" lastFinishedPulling="2025-10-05 08:59:36.515574333 +0000 UTC m=+7610.398200783" observedRunningTime="2025-10-05 08:59:37.544987235 +0000 UTC m=+7611.427613695" watchObservedRunningTime="2025-10-05 08:59:37.552586977 +0000 UTC m=+7611.435213437" Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.561652 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.596704 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-qhnds" podStartSLOduration=2.426174157 podStartE2EDuration="9.596689146s" podCreationTimestamp="2025-10-05 08:59:28 +0000 UTC" firstStartedPulling="2025-10-05 08:59:29.564137122 +0000 UTC m=+7603.446763582" lastFinishedPulling="2025-10-05 08:59:36.734652111 +0000 UTC m=+7610.617278571" observedRunningTime="2025-10-05 08:59:37.592519976 +0000 UTC m=+7611.475146426" watchObservedRunningTime="2025-10-05 08:59:37.596689146 +0000 UTC m=+7611.479315606" Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.639681 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-649566fff-bckjm" podStartSLOduration=2.519859451 podStartE2EDuration="9.639659895s" podCreationTimestamp="2025-10-05 08:59:28 +0000 UTC" firstStartedPulling="2025-10-05 08:59:29.395654545 +0000 UTC m=+7603.278281005" lastFinishedPulling="2025-10-05 08:59:36.515454989 +0000 UTC m=+7610.398081449" observedRunningTime="2025-10-05 08:59:37.624322609 +0000 UTC m=+7611.506949069" watchObservedRunningTime="2025-10-05 08:59:37.639659895 +0000 UTC m=+7611.522286355" Oct 05 08:59:37 crc kubenswrapper[4935]: I1005 08:59:37.654755 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" podStartSLOduration=2.5466471200000003 podStartE2EDuration="9.654740005s" podCreationTimestamp="2025-10-05 08:59:28 +0000 UTC" firstStartedPulling="2025-10-05 08:59:29.407326384 +0000 UTC m=+7603.289952834" lastFinishedPulling="2025-10-05 08:59:36.515419259 +0000 UTC m=+7610.398045719" observedRunningTime="2025-10-05 08:59:37.651427617 +0000 UTC m=+7611.534054077" watchObservedRunningTime="2025-10-05 08:59:37.654740005 +0000 UTC m=+7611.537366455" Oct 05 08:59:42 crc kubenswrapper[4935]: I1005 08:59:42.043133 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-3d00-account-create-hdc7z"] Oct 05 08:59:42 crc kubenswrapper[4935]: I1005 08:59:42.051338 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-3d00-account-create-hdc7z"] Oct 05 08:59:42 crc kubenswrapper[4935]: I1005 08:59:42.792537 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae0958d8-c9b8-4c10-b422-a66922f5c96b" path="/var/lib/kubelet/pods/ae0958d8-c9b8-4c10-b422-a66922f5c96b/volumes" Oct 05 08:59:44 crc kubenswrapper[4935]: I1005 08:59:44.302088 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:59:44 crc kubenswrapper[4935]: I1005 08:59:44.302497 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:59:48 crc kubenswrapper[4935]: I1005 08:59:48.796873 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-ztdwx" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.365050 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.365496 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" containerName="openstackclient" containerID="cri-o://3ff8e08c666a64c87ec949ce6136a4e8df7ced86690ff5502b8fbdba5138fcb4" gracePeriod=2 Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.375357 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.431939 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:51 crc kubenswrapper[4935]: E1005 08:59:51.432371 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" containerName="openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.432389 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" containerName="openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.432610 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" containerName="openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.433312 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.452791 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.469115 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" podUID="ecb82943-a9ff-44a6-a615-d5263d4c7e16" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.493611 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:51 crc kubenswrapper[4935]: E1005 08:59:51.494286 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-tn5tf openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="ecb82943-a9ff-44a6-a615-d5263d4c7e16" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.506114 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.530831 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.532101 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.554960 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.594786 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config-secret\") pod \"openstackclient\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.594869 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e58adf26-aff2-4a22-a7bb-47fb6664d949-openstack-config\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.594906 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwl5x\" (UniqueName: \"kubernetes.io/projected/e58adf26-aff2-4a22-a7bb-47fb6664d949-kube-api-access-bwl5x\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.594943 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config\") pod \"openstackclient\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.594964 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e58adf26-aff2-4a22-a7bb-47fb6664d949-openstack-config-secret\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.595006 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn5tf\" (UniqueName: \"kubernetes.io/projected/ecb82943-a9ff-44a6-a615-d5263d4c7e16-kube-api-access-tn5tf\") pod \"openstackclient\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.623581 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.625064 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.630291 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-q4bbs" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.642117 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.685685 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.695418 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ecb82943-a9ff-44a6-a615-d5263d4c7e16" podUID="e58adf26-aff2-4a22-a7bb-47fb6664d949" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.697504 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwl5x\" (UniqueName: \"kubernetes.io/projected/e58adf26-aff2-4a22-a7bb-47fb6664d949-kube-api-access-bwl5x\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.697539 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e58adf26-aff2-4a22-a7bb-47fb6664d949-openstack-config\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.697583 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config\") pod \"openstackclient\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.697606 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e58adf26-aff2-4a22-a7bb-47fb6664d949-openstack-config-secret\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.697651 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn5tf\" (UniqueName: \"kubernetes.io/projected/ecb82943-a9ff-44a6-a615-d5263d4c7e16-kube-api-access-tn5tf\") pod \"openstackclient\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.697719 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xskt\" (UniqueName: \"kubernetes.io/projected/71af6336-7272-441b-ba9e-a14b824af3c4-kube-api-access-5xskt\") pod \"kube-state-metrics-0\" (UID: \"71af6336-7272-441b-ba9e-a14b824af3c4\") " pod="openstack/kube-state-metrics-0" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.697752 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config-secret\") pod \"openstackclient\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.699426 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config\") pod \"openstackclient\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.700290 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e58adf26-aff2-4a22-a7bb-47fb6664d949-openstack-config\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: E1005 08:59:51.705175 4935 projected.go:194] Error preparing data for projected volume kube-api-access-tn5tf for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (ecb82943-a9ff-44a6-a615-d5263d4c7e16) does not match the UID in record. The object might have been deleted and then recreated Oct 05 08:59:51 crc kubenswrapper[4935]: E1005 08:59:51.705246 4935 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ecb82943-a9ff-44a6-a615-d5263d4c7e16-kube-api-access-tn5tf podName:ecb82943-a9ff-44a6-a615-d5263d4c7e16 nodeName:}" failed. No retries permitted until 2025-10-05 08:59:52.205226451 +0000 UTC m=+7626.087852901 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-tn5tf" (UniqueName: "kubernetes.io/projected/ecb82943-a9ff-44a6-a615-d5263d4c7e16-kube-api-access-tn5tf") pod "openstackclient" (UID: "ecb82943-a9ff-44a6-a615-d5263d4c7e16") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (ecb82943-a9ff-44a6-a615-d5263d4c7e16) does not match the UID in record. The object might have been deleted and then recreated Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.710228 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config-secret\") pod \"openstackclient\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.717877 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e58adf26-aff2-4a22-a7bb-47fb6664d949-openstack-config-secret\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.733844 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwl5x\" (UniqueName: \"kubernetes.io/projected/e58adf26-aff2-4a22-a7bb-47fb6664d949-kube-api-access-bwl5x\") pod \"openstackclient\" (UID: \"e58adf26-aff2-4a22-a7bb-47fb6664d949\") " pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.799529 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xskt\" (UniqueName: \"kubernetes.io/projected/71af6336-7272-441b-ba9e-a14b824af3c4-kube-api-access-5xskt\") pod \"kube-state-metrics-0\" (UID: \"71af6336-7272-441b-ba9e-a14b824af3c4\") " pod="openstack/kube-state-metrics-0" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.823835 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.825632 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xskt\" (UniqueName: \"kubernetes.io/projected/71af6336-7272-441b-ba9e-a14b824af3c4-kube-api-access-5xskt\") pod \"kube-state-metrics-0\" (UID: \"71af6336-7272-441b-ba9e-a14b824af3c4\") " pod="openstack/kube-state-metrics-0" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.855339 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.905578 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config\") pod \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.905807 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config-secret\") pod \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\" (UID: \"ecb82943-a9ff-44a6-a615-d5263d4c7e16\") " Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.906304 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tn5tf\" (UniqueName: \"kubernetes.io/projected/ecb82943-a9ff-44a6-a615-d5263d4c7e16-kube-api-access-tn5tf\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.906365 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ecb82943-a9ff-44a6-a615-d5263d4c7e16" (UID: "ecb82943-a9ff-44a6-a615-d5263d4c7e16"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:59:51 crc kubenswrapper[4935]: I1005 08:59:51.930050 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ecb82943-a9ff-44a6-a615-d5263d4c7e16" (UID: "ecb82943-a9ff-44a6-a615-d5263d4c7e16"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.007672 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.007697 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecb82943-a9ff-44a6-a615-d5263d4c7e16-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.116763 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.344902 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.350508 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.360197 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.360267 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.366215 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.367014 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-smpjg" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.419954 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.542920 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f5d04268-2dab-4d23-b434-b91cb01108f9-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.542987 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h8hf\" (UniqueName: \"kubernetes.io/projected/f5d04268-2dab-4d23-b434-b91cb01108f9-kube-api-access-2h8hf\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.543028 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f5d04268-2dab-4d23-b434-b91cb01108f9-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.543059 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/f5d04268-2dab-4d23-b434-b91cb01108f9-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.543089 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f5d04268-2dab-4d23-b434-b91cb01108f9-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.543109 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f5d04268-2dab-4d23-b434-b91cb01108f9-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.591925 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.645038 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f5d04268-2dab-4d23-b434-b91cb01108f9-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.645119 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h8hf\" (UniqueName: \"kubernetes.io/projected/f5d04268-2dab-4d23-b434-b91cb01108f9-kube-api-access-2h8hf\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.645158 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f5d04268-2dab-4d23-b434-b91cb01108f9-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.645188 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/f5d04268-2dab-4d23-b434-b91cb01108f9-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.645220 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f5d04268-2dab-4d23-b434-b91cb01108f9-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.645243 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f5d04268-2dab-4d23-b434-b91cb01108f9-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.651573 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/f5d04268-2dab-4d23-b434-b91cb01108f9-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.653096 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f5d04268-2dab-4d23-b434-b91cb01108f9-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.659262 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f5d04268-2dab-4d23-b434-b91cb01108f9-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.668503 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f5d04268-2dab-4d23-b434-b91cb01108f9-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.673817 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f5d04268-2dab-4d23-b434-b91cb01108f9-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.687722 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h8hf\" (UniqueName: \"kubernetes.io/projected/f5d04268-2dab-4d23-b434-b91cb01108f9-kube-api-access-2h8hf\") pod \"alertmanager-metric-storage-0\" (UID: \"f5d04268-2dab-4d23-b434-b91cb01108f9\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.711594 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.717929 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.720009 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e58adf26-aff2-4a22-a7bb-47fb6664d949","Type":"ContainerStarted","Data":"666661d0349ef504c8d98ff0829fe37cd6f87f4fb23a0a631141862ddbc99447"} Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.745278 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ecb82943-a9ff-44a6-a615-d5263d4c7e16" podUID="e58adf26-aff2-4a22-a7bb-47fb6664d949" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.857050 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb82943-a9ff-44a6-a615-d5263d4c7e16" path="/var/lib/kubelet/pods/ecb82943-a9ff-44a6-a615-d5263d4c7e16/volumes" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.908947 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.911499 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.915819 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.915984 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.917294 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.917511 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.917661 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mzm5m" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.917799 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.939981 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:59:52 crc kubenswrapper[4935]: W1005 08:59:52.958981 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71af6336_7272_441b_ba9e_a14b824af3c4.slice/crio-a47659d63094adcae0d077e90ddd5eafe86767916f3c62c82c37938446fbf37d WatchSource:0}: Error finding container a47659d63094adcae0d077e90ddd5eafe86767916f3c62c82c37938446fbf37d: Status 404 returned error can't find the container with id a47659d63094adcae0d077e90ddd5eafe86767916f3c62c82c37938446fbf37d Oct 05 08:59:52 crc kubenswrapper[4935]: I1005 08:59:52.962864 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.051355 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e18b6457-4bba-40d7-abad-577c4d251f45-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.051392 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-config\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.051421 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s8lz\" (UniqueName: \"kubernetes.io/projected/e18b6457-4bba-40d7-abad-577c4d251f45-kube-api-access-6s8lz\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.051469 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.051493 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2dc7e02a-edd8-4edb-846b-e893ac3a3bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2dc7e02a-edd8-4edb-846b-e893ac3a3bdb\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.051722 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e18b6457-4bba-40d7-abad-577c4d251f45-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.051795 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e18b6457-4bba-40d7-abad-577c4d251f45-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.051874 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.155452 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s8lz\" (UniqueName: \"kubernetes.io/projected/e18b6457-4bba-40d7-abad-577c4d251f45-kube-api-access-6s8lz\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.155802 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.155834 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2dc7e02a-edd8-4edb-846b-e893ac3a3bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2dc7e02a-edd8-4edb-846b-e893ac3a3bdb\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.155905 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e18b6457-4bba-40d7-abad-577c4d251f45-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.155935 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e18b6457-4bba-40d7-abad-577c4d251f45-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.155966 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.156034 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e18b6457-4bba-40d7-abad-577c4d251f45-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.156057 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-config\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.157396 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e18b6457-4bba-40d7-abad-577c4d251f45-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.176319 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.176482 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.187067 4935 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.187109 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2dc7e02a-edd8-4edb-846b-e893ac3a3bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2dc7e02a-edd8-4edb-846b-e893ac3a3bdb\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3bd7be0540566392db2806b239fb424125eb6c6a9f3bb17d9556a29383dec176/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.188342 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e18b6457-4bba-40d7-abad-577c4d251f45-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.196048 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e18b6457-4bba-40d7-abad-577c4d251f45-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.196077 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s8lz\" (UniqueName: \"kubernetes.io/projected/e18b6457-4bba-40d7-abad-577c4d251f45-kube-api-access-6s8lz\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.201575 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e18b6457-4bba-40d7-abad-577c4d251f45-config\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.323450 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2dc7e02a-edd8-4edb-846b-e893ac3a3bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2dc7e02a-edd8-4edb-846b-e893ac3a3bdb\") pod \"prometheus-metric-storage-0\" (UID: \"e18b6457-4bba-40d7-abad-577c4d251f45\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.351311 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.548821 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.731139 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"71af6336-7272-441b-ba9e-a14b824af3c4","Type":"ContainerStarted","Data":"a47659d63094adcae0d077e90ddd5eafe86767916f3c62c82c37938446fbf37d"} Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.736711 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e58adf26-aff2-4a22-a7bb-47fb6664d949","Type":"ContainerStarted","Data":"9c3209746644aa48bf3718f8f9d731d128c6d4b6d63c9d95e3c8f05704f997c1"} Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.739531 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"f5d04268-2dab-4d23-b434-b91cb01108f9","Type":"ContainerStarted","Data":"7bc52a1b7fd74ba3d6027487260da5f02a70fe7c301a4724a93ca3318c1113a4"} Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.742877 4935 generic.go:334] "Generic (PLEG): container finished" podID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" containerID="3ff8e08c666a64c87ec949ce6136a4e8df7ced86690ff5502b8fbdba5138fcb4" exitCode=137 Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.752538 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.752517101 podStartE2EDuration="2.752517101s" podCreationTimestamp="2025-10-05 08:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:59:53.749346447 +0000 UTC m=+7627.631972907" watchObservedRunningTime="2025-10-05 08:59:53.752517101 +0000 UTC m=+7627.635143561" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.839557 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.975233 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.975352 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config\") pod \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.975676 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config-secret\") pod \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.975837 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcs8c\" (UniqueName: \"kubernetes.io/projected/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-kube-api-access-fcs8c\") pod \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\" (UID: \"5810bf2f-6c1e-465b-9ad4-a17ac78a0a56\") " Oct 05 08:59:53 crc kubenswrapper[4935]: I1005 08:59:53.981251 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-kube-api-access-fcs8c" (OuterVolumeSpecName: "kube-api-access-fcs8c") pod "5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" (UID: "5810bf2f-6c1e-465b-9ad4-a17ac78a0a56"). InnerVolumeSpecName "kube-api-access-fcs8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.003315 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" (UID: "5810bf2f-6c1e-465b-9ad4-a17ac78a0a56"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.045118 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" (UID: "5810bf2f-6c1e-465b-9ad4-a17ac78a0a56"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.078051 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcs8c\" (UniqueName: \"kubernetes.io/projected/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-kube-api-access-fcs8c\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.078326 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.078438 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.754761 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.754819 4935 scope.go:117] "RemoveContainer" containerID="3ff8e08c666a64c87ec949ce6136a4e8df7ced86690ff5502b8fbdba5138fcb4" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.757653 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"71af6336-7272-441b-ba9e-a14b824af3c4","Type":"ContainerStarted","Data":"eedefb12da039230893b2f46fe17089fb892157bd8d59d59053a9040f9469615"} Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.758201 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.761214 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e18b6457-4bba-40d7-abad-577c4d251f45","Type":"ContainerStarted","Data":"4a1f4391723bb1c03f156da91b00f6fdcf7950c8c7339002c9401eb3084783ca"} Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.776401 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.235216858 podStartE2EDuration="3.776378775s" podCreationTimestamp="2025-10-05 08:59:51 +0000 UTC" firstStartedPulling="2025-10-05 08:59:52.977527734 +0000 UTC m=+7626.860154194" lastFinishedPulling="2025-10-05 08:59:53.518689661 +0000 UTC m=+7627.401316111" observedRunningTime="2025-10-05 08:59:54.773737495 +0000 UTC m=+7628.656363965" watchObservedRunningTime="2025-10-05 08:59:54.776378775 +0000 UTC m=+7628.659005235" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.782300 4935 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" podUID="e58adf26-aff2-4a22-a7bb-47fb6664d949" Oct 05 08:59:54 crc kubenswrapper[4935]: I1005 08:59:54.791105 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5810bf2f-6c1e-465b-9ad4-a17ac78a0a56" path="/var/lib/kubelet/pods/5810bf2f-6c1e-465b-9ad4-a17ac78a0a56/volumes" Oct 05 08:59:54 crc kubenswrapper[4935]: E1005 08:59:54.919824 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5810bf2f_6c1e_465b_9ad4_a17ac78a0a56.slice\": RecentStats: unable to find data in memory cache]" Oct 05 08:59:58 crc kubenswrapper[4935]: I1005 08:59:58.808283 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"f5d04268-2dab-4d23-b434-b91cb01108f9","Type":"ContainerStarted","Data":"77b5415f7812a0092e35c25622e89f074f747376a4050155850cd7699169e39a"} Oct 05 08:59:59 crc kubenswrapper[4935]: I1005 08:59:59.836436 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e18b6457-4bba-40d7-abad-577c4d251f45","Type":"ContainerStarted","Data":"ae1974cb9926c1d194f18fb1fc77d843cb2e4c6befd169b8c736727603a38a96"} Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.139842 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz"] Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.141181 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.143618 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.144021 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.161367 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz"] Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.212632 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-config-volume\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.212828 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-secret-volume\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.213037 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zf77\" (UniqueName: \"kubernetes.io/projected/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-kube-api-access-4zf77\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.314585 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-secret-volume\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.314718 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zf77\" (UniqueName: \"kubernetes.io/projected/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-kube-api-access-4zf77\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.314794 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-config-volume\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.315927 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-config-volume\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.325166 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-secret-volume\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.333680 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zf77\" (UniqueName: \"kubernetes.io/projected/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-kube-api-access-4zf77\") pod \"collect-profiles-29327580-r4hqz\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.480767 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.837017 4935 scope.go:117] "RemoveContainer" containerID="7a5141487b7219c38654c0be6915513a17cd06b9d7b917b24d2fdb28949e18e2" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.859704 4935 scope.go:117] "RemoveContainer" containerID="3eeb54cb99164ce98b1ba3a478e7c2bf72a7b055ccaf933aefdc8b4e9da97cff" Oct 05 09:00:00 crc kubenswrapper[4935]: I1005 09:00:00.945920 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz"] Oct 05 09:00:01 crc kubenswrapper[4935]: I1005 09:00:01.856986 4935 generic.go:334] "Generic (PLEG): container finished" podID="deae3a78-35c4-4fd6-bf16-b2cda1460dd7" containerID="3d893bc44e401e3a801a1a454418d7deb79e1f7be7eebcd0d8b93cfb56bae961" exitCode=0 Oct 05 09:00:01 crc kubenswrapper[4935]: I1005 09:00:01.857276 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" event={"ID":"deae3a78-35c4-4fd6-bf16-b2cda1460dd7","Type":"ContainerDied","Data":"3d893bc44e401e3a801a1a454418d7deb79e1f7be7eebcd0d8b93cfb56bae961"} Oct 05 09:00:01 crc kubenswrapper[4935]: I1005 09:00:01.857305 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" event={"ID":"deae3a78-35c4-4fd6-bf16-b2cda1460dd7","Type":"ContainerStarted","Data":"7e4539a7b129accdb493f0c0e103843eed9974dcd2ca7ee9585ec3d39cde687b"} Oct 05 09:00:02 crc kubenswrapper[4935]: I1005 09:00:02.120876 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.294800 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.487368 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zf77\" (UniqueName: \"kubernetes.io/projected/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-kube-api-access-4zf77\") pod \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.487464 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-config-volume\") pod \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.487506 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-secret-volume\") pod \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\" (UID: \"deae3a78-35c4-4fd6-bf16-b2cda1460dd7\") " Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.488310 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-config-volume" (OuterVolumeSpecName: "config-volume") pod "deae3a78-35c4-4fd6-bf16-b2cda1460dd7" (UID: "deae3a78-35c4-4fd6-bf16-b2cda1460dd7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.495086 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "deae3a78-35c4-4fd6-bf16-b2cda1460dd7" (UID: "deae3a78-35c4-4fd6-bf16-b2cda1460dd7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.495172 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-kube-api-access-4zf77" (OuterVolumeSpecName: "kube-api-access-4zf77") pod "deae3a78-35c4-4fd6-bf16-b2cda1460dd7" (UID: "deae3a78-35c4-4fd6-bf16-b2cda1460dd7"). InnerVolumeSpecName "kube-api-access-4zf77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.589803 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zf77\" (UniqueName: \"kubernetes.io/projected/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-kube-api-access-4zf77\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.589848 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.589863 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/deae3a78-35c4-4fd6-bf16-b2cda1460dd7-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.876105 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" event={"ID":"deae3a78-35c4-4fd6-bf16-b2cda1460dd7","Type":"ContainerDied","Data":"7e4539a7b129accdb493f0c0e103843eed9974dcd2ca7ee9585ec3d39cde687b"} Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.876420 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e4539a7b129accdb493f0c0e103843eed9974dcd2ca7ee9585ec3d39cde687b" Oct 05 09:00:03 crc kubenswrapper[4935]: I1005 09:00:03.876166 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz" Oct 05 09:00:04 crc kubenswrapper[4935]: I1005 09:00:04.381343 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx"] Oct 05 09:00:04 crc kubenswrapper[4935]: I1005 09:00:04.392569 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-7xjqx"] Oct 05 09:00:04 crc kubenswrapper[4935]: I1005 09:00:04.791832 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38c72e22-c292-4f2a-99ac-810745491cf4" path="/var/lib/kubelet/pods/38c72e22-c292-4f2a-99ac-810745491cf4/volumes" Oct 05 09:00:05 crc kubenswrapper[4935]: I1005 09:00:05.059206 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-rqppz"] Oct 05 09:00:05 crc kubenswrapper[4935]: I1005 09:00:05.084496 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-rqppz"] Oct 05 09:00:06 crc kubenswrapper[4935]: I1005 09:00:06.809220 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0657582d-ba22-485a-9964-efb3e4b41fe7" path="/var/lib/kubelet/pods/0657582d-ba22-485a-9964-efb3e4b41fe7/volumes" Oct 05 09:00:06 crc kubenswrapper[4935]: I1005 09:00:06.918040 4935 generic.go:334] "Generic (PLEG): container finished" podID="f5d04268-2dab-4d23-b434-b91cb01108f9" containerID="77b5415f7812a0092e35c25622e89f074f747376a4050155850cd7699169e39a" exitCode=0 Oct 05 09:00:06 crc kubenswrapper[4935]: I1005 09:00:06.918154 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"f5d04268-2dab-4d23-b434-b91cb01108f9","Type":"ContainerDied","Data":"77b5415f7812a0092e35c25622e89f074f747376a4050155850cd7699169e39a"} Oct 05 09:00:08 crc kubenswrapper[4935]: I1005 09:00:08.936282 4935 generic.go:334] "Generic (PLEG): container finished" podID="e18b6457-4bba-40d7-abad-577c4d251f45" containerID="ae1974cb9926c1d194f18fb1fc77d843cb2e4c6befd169b8c736727603a38a96" exitCode=0 Oct 05 09:00:08 crc kubenswrapper[4935]: I1005 09:00:08.936435 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e18b6457-4bba-40d7-abad-577c4d251f45","Type":"ContainerDied","Data":"ae1974cb9926c1d194f18fb1fc77d843cb2e4c6befd169b8c736727603a38a96"} Oct 05 09:00:10 crc kubenswrapper[4935]: I1005 09:00:10.959949 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"f5d04268-2dab-4d23-b434-b91cb01108f9","Type":"ContainerStarted","Data":"c01cdfa923fdbf48cdc706f0994828623ce42626abeec541b3ef68b878f1fc72"} Oct 05 09:00:13 crc kubenswrapper[4935]: I1005 09:00:13.989379 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e18b6457-4bba-40d7-abad-577c4d251f45","Type":"ContainerStarted","Data":"39098f8355b32d8e23c1cba91bab4bed88912e73aa76b3d585af9ccc146d5a60"} Oct 05 09:00:13 crc kubenswrapper[4935]: I1005 09:00:13.992517 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"f5d04268-2dab-4d23-b434-b91cb01108f9","Type":"ContainerStarted","Data":"c9cc2fb9eae52333f4d897c6aad5e55fc96811bf6ef2d2f66ca267ada833f21f"} Oct 05 09:00:13 crc kubenswrapper[4935]: I1005 09:00:13.992820 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 05 09:00:13 crc kubenswrapper[4935]: I1005 09:00:13.996647 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 05 09:00:14 crc kubenswrapper[4935]: I1005 09:00:14.029220 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.667325093 podStartE2EDuration="22.029200538s" podCreationTimestamp="2025-10-05 08:59:52 +0000 UTC" firstStartedPulling="2025-10-05 08:59:53.573197297 +0000 UTC m=+7627.455823757" lastFinishedPulling="2025-10-05 09:00:09.935072722 +0000 UTC m=+7643.817699202" observedRunningTime="2025-10-05 09:00:14.019875041 +0000 UTC m=+7647.902501531" watchObservedRunningTime="2025-10-05 09:00:14.029200538 +0000 UTC m=+7647.911827008" Oct 05 09:00:14 crc kubenswrapper[4935]: I1005 09:00:14.289396 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:00:14 crc kubenswrapper[4935]: I1005 09:00:14.289464 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:00:17 crc kubenswrapper[4935]: I1005 09:00:17.029190 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e18b6457-4bba-40d7-abad-577c4d251f45","Type":"ContainerStarted","Data":"435fcbb7b1b5a526f445ee7f90e80de5bfbd41949330798f04e86b64a57db531"} Oct 05 09:00:21 crc kubenswrapper[4935]: I1005 09:00:21.097795 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e18b6457-4bba-40d7-abad-577c4d251f45","Type":"ContainerStarted","Data":"3fefc40d49fe8f8bad223096e8b98fe94ddecdce5b9be69157eab60b3e72321f"} Oct 05 09:00:21 crc kubenswrapper[4935]: I1005 09:00:21.145508 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=4.206937371 podStartE2EDuration="30.1454832s" podCreationTimestamp="2025-10-05 08:59:51 +0000 UTC" firstStartedPulling="2025-10-05 08:59:53.981229204 +0000 UTC m=+7627.863855664" lastFinishedPulling="2025-10-05 09:00:19.919775023 +0000 UTC m=+7653.802401493" observedRunningTime="2025-10-05 09:00:21.136578444 +0000 UTC m=+7655.019204934" watchObservedRunningTime="2025-10-05 09:00:21.1454832 +0000 UTC m=+7655.028109680" Oct 05 09:00:23 crc kubenswrapper[4935]: I1005 09:00:23.352359 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 05 09:00:23 crc kubenswrapper[4935]: I1005 09:00:23.352768 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 05 09:00:23 crc kubenswrapper[4935]: I1005 09:00:23.355121 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 05 09:00:24 crc kubenswrapper[4935]: I1005 09:00:24.129164 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.240552 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:00:25 crc kubenswrapper[4935]: E1005 09:00:25.243580 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deae3a78-35c4-4fd6-bf16-b2cda1460dd7" containerName="collect-profiles" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.243609 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="deae3a78-35c4-4fd6-bf16-b2cda1460dd7" containerName="collect-profiles" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.243807 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="deae3a78-35c4-4fd6-bf16-b2cda1460dd7" containerName="collect-profiles" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.245699 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.247752 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.248036 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.268363 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.343445 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.343493 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.343604 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-log-httpd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.343645 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-run-httpd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.343701 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-scripts\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.343748 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp2hd\" (UniqueName: \"kubernetes.io/projected/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-kube-api-access-vp2hd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.343783 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-config-data\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.445592 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-scripts\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.445648 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp2hd\" (UniqueName: \"kubernetes.io/projected/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-kube-api-access-vp2hd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.445680 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-config-data\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.445745 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.445762 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.445811 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-log-httpd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.445841 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-run-httpd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.446207 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-run-httpd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.446809 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-log-httpd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.453649 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.453792 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-scripts\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.454075 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-config-data\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.463504 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.464053 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp2hd\" (UniqueName: \"kubernetes.io/projected/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-kube-api-access-vp2hd\") pod \"ceilometer-0\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " pod="openstack/ceilometer-0" Oct 05 09:00:25 crc kubenswrapper[4935]: I1005 09:00:25.590497 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:00:26 crc kubenswrapper[4935]: I1005 09:00:26.181925 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:00:27 crc kubenswrapper[4935]: I1005 09:00:27.160396 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerStarted","Data":"9069cc1d9e934e88037fbaf3b8c7b9eeb6151755dfb88a46f123ce742d0ea3e2"} Oct 05 09:00:31 crc kubenswrapper[4935]: I1005 09:00:31.198887 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerStarted","Data":"b0f379a800d47443b8e9e414ae37a0d586a6e0469fd7b7ff86350d154bb6bbb7"} Oct 05 09:00:32 crc kubenswrapper[4935]: I1005 09:00:32.216962 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerStarted","Data":"8be51b3e4f763e7715afeb82eee38445aaca2dc6740b1be143f7d4be727101aa"} Oct 05 09:00:34 crc kubenswrapper[4935]: I1005 09:00:34.252469 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerStarted","Data":"32c2df667e2bf45da5733e1a0fe2831cc1607a2b6b71f7b0b75a3d97463fd1bc"} Oct 05 09:00:35 crc kubenswrapper[4935]: I1005 09:00:35.272110 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerStarted","Data":"9a10344778c84fa6aee5f7c67d6fefdf14ffe8c59ff9c56e68d7e69dfb61452b"} Oct 05 09:00:35 crc kubenswrapper[4935]: I1005 09:00:35.272503 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 09:00:35 crc kubenswrapper[4935]: I1005 09:00:35.299068 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8614510050000002 podStartE2EDuration="10.299054159s" podCreationTimestamp="2025-10-05 09:00:25 +0000 UTC" firstStartedPulling="2025-10-05 09:00:26.214057361 +0000 UTC m=+7660.096683821" lastFinishedPulling="2025-10-05 09:00:34.651660485 +0000 UTC m=+7668.534286975" observedRunningTime="2025-10-05 09:00:35.298745421 +0000 UTC m=+7669.181371891" watchObservedRunningTime="2025-10-05 09:00:35.299054159 +0000 UTC m=+7669.181680619" Oct 05 09:00:38 crc kubenswrapper[4935]: I1005 09:00:38.232378 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-4bqbw"] Oct 05 09:00:38 crc kubenswrapper[4935]: I1005 09:00:38.234477 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4bqbw" Oct 05 09:00:38 crc kubenswrapper[4935]: I1005 09:00:38.252557 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-4bqbw"] Oct 05 09:00:38 crc kubenswrapper[4935]: I1005 09:00:38.316537 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp7v8\" (UniqueName: \"kubernetes.io/projected/0249f885-bcbd-45f3-bbb6-b1285f4b47d0-kube-api-access-kp7v8\") pod \"aodh-db-create-4bqbw\" (UID: \"0249f885-bcbd-45f3-bbb6-b1285f4b47d0\") " pod="openstack/aodh-db-create-4bqbw" Oct 05 09:00:38 crc kubenswrapper[4935]: I1005 09:00:38.418517 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp7v8\" (UniqueName: \"kubernetes.io/projected/0249f885-bcbd-45f3-bbb6-b1285f4b47d0-kube-api-access-kp7v8\") pod \"aodh-db-create-4bqbw\" (UID: \"0249f885-bcbd-45f3-bbb6-b1285f4b47d0\") " pod="openstack/aodh-db-create-4bqbw" Oct 05 09:00:38 crc kubenswrapper[4935]: I1005 09:00:38.453119 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp7v8\" (UniqueName: \"kubernetes.io/projected/0249f885-bcbd-45f3-bbb6-b1285f4b47d0-kube-api-access-kp7v8\") pod \"aodh-db-create-4bqbw\" (UID: \"0249f885-bcbd-45f3-bbb6-b1285f4b47d0\") " pod="openstack/aodh-db-create-4bqbw" Oct 05 09:00:38 crc kubenswrapper[4935]: I1005 09:00:38.571073 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4bqbw" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.028665 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-rjlzp"] Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.037671 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-rjlzp"] Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.058192 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-4bqbw"] Oct 05 09:00:39 crc kubenswrapper[4935]: W1005 09:00:39.071118 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0249f885_bcbd_45f3_bbb6_b1285f4b47d0.slice/crio-20d04208d57b6e73feb96f2a754fc51bee29e93713f38d160275754e891a5dba WatchSource:0}: Error finding container 20d04208d57b6e73feb96f2a754fc51bee29e93713f38d160275754e891a5dba: Status 404 returned error can't find the container with id 20d04208d57b6e73feb96f2a754fc51bee29e93713f38d160275754e891a5dba Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.143617 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9wjl2"] Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.145823 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.169892 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wjl2"] Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.234646 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-catalog-content\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.234692 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrjp6\" (UniqueName: \"kubernetes.io/projected/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-kube-api-access-vrjp6\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.234772 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-utilities\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.314685 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4bqbw" event={"ID":"0249f885-bcbd-45f3-bbb6-b1285f4b47d0","Type":"ContainerStarted","Data":"61ad34ae00274c7a1357ebf1f9f69d5bfb9c5d015b8c87a3a29465393b473ace"} Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.314739 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4bqbw" event={"ID":"0249f885-bcbd-45f3-bbb6-b1285f4b47d0","Type":"ContainerStarted","Data":"20d04208d57b6e73feb96f2a754fc51bee29e93713f38d160275754e891a5dba"} Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.333921 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-4bqbw" podStartSLOduration=1.333881073 podStartE2EDuration="1.333881073s" podCreationTimestamp="2025-10-05 09:00:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:00:39.32657693 +0000 UTC m=+7673.209203390" watchObservedRunningTime="2025-10-05 09:00:39.333881073 +0000 UTC m=+7673.216507523" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.336725 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-catalog-content\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.336769 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrjp6\" (UniqueName: \"kubernetes.io/projected/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-kube-api-access-vrjp6\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.336871 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-utilities\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.337351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-catalog-content\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.337361 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-utilities\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.358038 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrjp6\" (UniqueName: \"kubernetes.io/projected/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-kube-api-access-vrjp6\") pod \"community-operators-9wjl2\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:39 crc kubenswrapper[4935]: I1005 09:00:39.523837 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:40 crc kubenswrapper[4935]: I1005 09:00:40.083751 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wjl2"] Oct 05 09:00:40 crc kubenswrapper[4935]: I1005 09:00:40.328177 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wjl2" event={"ID":"076cd1d5-0f9d-4497-9abf-de7dadd41ca9","Type":"ContainerStarted","Data":"3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423"} Oct 05 09:00:40 crc kubenswrapper[4935]: I1005 09:00:40.328242 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wjl2" event={"ID":"076cd1d5-0f9d-4497-9abf-de7dadd41ca9","Type":"ContainerStarted","Data":"dd9cd45db54ba0b3b4c7776048d0b9825d5a770a375d50d64f83fbc38010969b"} Oct 05 09:00:40 crc kubenswrapper[4935]: I1005 09:00:40.330537 4935 generic.go:334] "Generic (PLEG): container finished" podID="0249f885-bcbd-45f3-bbb6-b1285f4b47d0" containerID="61ad34ae00274c7a1357ebf1f9f69d5bfb9c5d015b8c87a3a29465393b473ace" exitCode=0 Oct 05 09:00:40 crc kubenswrapper[4935]: I1005 09:00:40.330592 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4bqbw" event={"ID":"0249f885-bcbd-45f3-bbb6-b1285f4b47d0","Type":"ContainerDied","Data":"61ad34ae00274c7a1357ebf1f9f69d5bfb9c5d015b8c87a3a29465393b473ace"} Oct 05 09:00:40 crc kubenswrapper[4935]: I1005 09:00:40.796152 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c82d411-42ad-40d0-a507-8805af16350f" path="/var/lib/kubelet/pods/7c82d411-42ad-40d0-a507-8805af16350f/volumes" Oct 05 09:00:41 crc kubenswrapper[4935]: I1005 09:00:41.355028 4935 generic.go:334] "Generic (PLEG): container finished" podID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerID="3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423" exitCode=0 Oct 05 09:00:41 crc kubenswrapper[4935]: I1005 09:00:41.355159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wjl2" event={"ID":"076cd1d5-0f9d-4497-9abf-de7dadd41ca9","Type":"ContainerDied","Data":"3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423"} Oct 05 09:00:41 crc kubenswrapper[4935]: I1005 09:00:41.355323 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wjl2" event={"ID":"076cd1d5-0f9d-4497-9abf-de7dadd41ca9","Type":"ContainerStarted","Data":"eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838"} Oct 05 09:00:41 crc kubenswrapper[4935]: I1005 09:00:41.756478 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4bqbw" Oct 05 09:00:41 crc kubenswrapper[4935]: I1005 09:00:41.898276 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp7v8\" (UniqueName: \"kubernetes.io/projected/0249f885-bcbd-45f3-bbb6-b1285f4b47d0-kube-api-access-kp7v8\") pod \"0249f885-bcbd-45f3-bbb6-b1285f4b47d0\" (UID: \"0249f885-bcbd-45f3-bbb6-b1285f4b47d0\") " Oct 05 09:00:41 crc kubenswrapper[4935]: I1005 09:00:41.904582 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0249f885-bcbd-45f3-bbb6-b1285f4b47d0-kube-api-access-kp7v8" (OuterVolumeSpecName: "kube-api-access-kp7v8") pod "0249f885-bcbd-45f3-bbb6-b1285f4b47d0" (UID: "0249f885-bcbd-45f3-bbb6-b1285f4b47d0"). InnerVolumeSpecName "kube-api-access-kp7v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:00:42 crc kubenswrapper[4935]: I1005 09:00:42.000991 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp7v8\" (UniqueName: \"kubernetes.io/projected/0249f885-bcbd-45f3-bbb6-b1285f4b47d0-kube-api-access-kp7v8\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:42 crc kubenswrapper[4935]: I1005 09:00:42.372752 4935 generic.go:334] "Generic (PLEG): container finished" podID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerID="eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838" exitCode=0 Oct 05 09:00:42 crc kubenswrapper[4935]: I1005 09:00:42.372835 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wjl2" event={"ID":"076cd1d5-0f9d-4497-9abf-de7dadd41ca9","Type":"ContainerDied","Data":"eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838"} Oct 05 09:00:42 crc kubenswrapper[4935]: I1005 09:00:42.380174 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4bqbw" event={"ID":"0249f885-bcbd-45f3-bbb6-b1285f4b47d0","Type":"ContainerDied","Data":"20d04208d57b6e73feb96f2a754fc51bee29e93713f38d160275754e891a5dba"} Oct 05 09:00:42 crc kubenswrapper[4935]: I1005 09:00:42.380237 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20d04208d57b6e73feb96f2a754fc51bee29e93713f38d160275754e891a5dba" Oct 05 09:00:42 crc kubenswrapper[4935]: I1005 09:00:42.380320 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4bqbw" Oct 05 09:00:44 crc kubenswrapper[4935]: I1005 09:00:44.289576 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:00:44 crc kubenswrapper[4935]: I1005 09:00:44.290718 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:00:44 crc kubenswrapper[4935]: I1005 09:00:44.290864 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:00:44 crc kubenswrapper[4935]: I1005 09:00:44.291867 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:00:44 crc kubenswrapper[4935]: I1005 09:00:44.292017 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" gracePeriod=600 Oct 05 09:00:44 crc kubenswrapper[4935]: I1005 09:00:44.403740 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wjl2" event={"ID":"076cd1d5-0f9d-4497-9abf-de7dadd41ca9","Type":"ContainerStarted","Data":"29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206"} Oct 05 09:00:44 crc kubenswrapper[4935]: E1005 09:00:44.449647 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:00:44 crc kubenswrapper[4935]: I1005 09:00:44.488068 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9wjl2" podStartSLOduration=2.6256974939999997 podStartE2EDuration="5.488045073s" podCreationTimestamp="2025-10-05 09:00:39 +0000 UTC" firstStartedPulling="2025-10-05 09:00:40.331579675 +0000 UTC m=+7674.214206145" lastFinishedPulling="2025-10-05 09:00:43.193927224 +0000 UTC m=+7677.076553724" observedRunningTime="2025-10-05 09:00:44.478257054 +0000 UTC m=+7678.360883534" watchObservedRunningTime="2025-10-05 09:00:44.488045073 +0000 UTC m=+7678.370671543" Oct 05 09:00:45 crc kubenswrapper[4935]: I1005 09:00:45.414539 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" exitCode=0 Oct 05 09:00:45 crc kubenswrapper[4935]: I1005 09:00:45.414609 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652"} Oct 05 09:00:45 crc kubenswrapper[4935]: I1005 09:00:45.414979 4935 scope.go:117] "RemoveContainer" containerID="4f78b54294a258ac737523e93c508737226a48f0df445e1a69f2784b25cc51ab" Oct 05 09:00:45 crc kubenswrapper[4935]: I1005 09:00:45.415767 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:00:45 crc kubenswrapper[4935]: E1005 09:00:45.416255 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.338240 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-35a0-account-create-fhstx"] Oct 05 09:00:48 crc kubenswrapper[4935]: E1005 09:00:48.339437 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0249f885-bcbd-45f3-bbb6-b1285f4b47d0" containerName="mariadb-database-create" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.339477 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0249f885-bcbd-45f3-bbb6-b1285f4b47d0" containerName="mariadb-database-create" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.339848 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0249f885-bcbd-45f3-bbb6-b1285f4b47d0" containerName="mariadb-database-create" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.340984 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-35a0-account-create-fhstx" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.344459 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.349265 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-35a0-account-create-fhstx"] Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.440535 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc7xg\" (UniqueName: \"kubernetes.io/projected/0b6dceca-640d-4af2-845d-48741604e927-kube-api-access-lc7xg\") pod \"aodh-35a0-account-create-fhstx\" (UID: \"0b6dceca-640d-4af2-845d-48741604e927\") " pod="openstack/aodh-35a0-account-create-fhstx" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.543010 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc7xg\" (UniqueName: \"kubernetes.io/projected/0b6dceca-640d-4af2-845d-48741604e927-kube-api-access-lc7xg\") pod \"aodh-35a0-account-create-fhstx\" (UID: \"0b6dceca-640d-4af2-845d-48741604e927\") " pod="openstack/aodh-35a0-account-create-fhstx" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.567677 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc7xg\" (UniqueName: \"kubernetes.io/projected/0b6dceca-640d-4af2-845d-48741604e927-kube-api-access-lc7xg\") pod \"aodh-35a0-account-create-fhstx\" (UID: \"0b6dceca-640d-4af2-845d-48741604e927\") " pod="openstack/aodh-35a0-account-create-fhstx" Oct 05 09:00:48 crc kubenswrapper[4935]: I1005 09:00:48.669206 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-35a0-account-create-fhstx" Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.034385 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-acdd-account-create-8qtbg"] Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.047172 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-acdd-account-create-8qtbg"] Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.130023 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-35a0-account-create-fhstx"] Oct 05 09:00:49 crc kubenswrapper[4935]: W1005 09:00:49.138486 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b6dceca_640d_4af2_845d_48741604e927.slice/crio-23f47c5a435df1d6ac7dbd336db81c14580fb5bf1f7c3404cd06aaf0a815636f WatchSource:0}: Error finding container 23f47c5a435df1d6ac7dbd336db81c14580fb5bf1f7c3404cd06aaf0a815636f: Status 404 returned error can't find the container with id 23f47c5a435df1d6ac7dbd336db81c14580fb5bf1f7c3404cd06aaf0a815636f Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.459596 4935 generic.go:334] "Generic (PLEG): container finished" podID="0b6dceca-640d-4af2-845d-48741604e927" containerID="2128de18a6b393434ef20a05a2786e15dce2a908b0b2f84319d87254057f62ee" exitCode=0 Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.459644 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-35a0-account-create-fhstx" event={"ID":"0b6dceca-640d-4af2-845d-48741604e927","Type":"ContainerDied","Data":"2128de18a6b393434ef20a05a2786e15dce2a908b0b2f84319d87254057f62ee"} Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.459674 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-35a0-account-create-fhstx" event={"ID":"0b6dceca-640d-4af2-845d-48741604e927","Type":"ContainerStarted","Data":"23f47c5a435df1d6ac7dbd336db81c14580fb5bf1f7c3404cd06aaf0a815636f"} Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.524615 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.524981 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:49 crc kubenswrapper[4935]: I1005 09:00:49.594326 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:50 crc kubenswrapper[4935]: I1005 09:00:50.528394 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:50 crc kubenswrapper[4935]: I1005 09:00:50.595965 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wjl2"] Oct 05 09:00:50 crc kubenswrapper[4935]: I1005 09:00:50.790952 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35816157-f3f5-4908-9bb0-0480b2d8970e" path="/var/lib/kubelet/pods/35816157-f3f5-4908-9bb0-0480b2d8970e/volumes" Oct 05 09:00:50 crc kubenswrapper[4935]: I1005 09:00:50.875827 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-35a0-account-create-fhstx" Oct 05 09:00:51 crc kubenswrapper[4935]: I1005 09:00:51.003243 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc7xg\" (UniqueName: \"kubernetes.io/projected/0b6dceca-640d-4af2-845d-48741604e927-kube-api-access-lc7xg\") pod \"0b6dceca-640d-4af2-845d-48741604e927\" (UID: \"0b6dceca-640d-4af2-845d-48741604e927\") " Oct 05 09:00:51 crc kubenswrapper[4935]: I1005 09:00:51.009420 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b6dceca-640d-4af2-845d-48741604e927-kube-api-access-lc7xg" (OuterVolumeSpecName: "kube-api-access-lc7xg") pod "0b6dceca-640d-4af2-845d-48741604e927" (UID: "0b6dceca-640d-4af2-845d-48741604e927"). InnerVolumeSpecName "kube-api-access-lc7xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:00:51 crc kubenswrapper[4935]: I1005 09:00:51.107084 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc7xg\" (UniqueName: \"kubernetes.io/projected/0b6dceca-640d-4af2-845d-48741604e927-kube-api-access-lc7xg\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:51 crc kubenswrapper[4935]: I1005 09:00:51.482107 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-35a0-account-create-fhstx" event={"ID":"0b6dceca-640d-4af2-845d-48741604e927","Type":"ContainerDied","Data":"23f47c5a435df1d6ac7dbd336db81c14580fb5bf1f7c3404cd06aaf0a815636f"} Oct 05 09:00:51 crc kubenswrapper[4935]: I1005 09:00:51.482146 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-35a0-account-create-fhstx" Oct 05 09:00:51 crc kubenswrapper[4935]: I1005 09:00:51.482168 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23f47c5a435df1d6ac7dbd336db81c14580fb5bf1f7c3404cd06aaf0a815636f" Oct 05 09:00:52 crc kubenswrapper[4935]: I1005 09:00:52.491425 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9wjl2" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerName="registry-server" containerID="cri-o://29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206" gracePeriod=2 Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.061792 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.150158 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-utilities\") pod \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.150198 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-catalog-content\") pod \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.150415 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrjp6\" (UniqueName: \"kubernetes.io/projected/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-kube-api-access-vrjp6\") pod \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\" (UID: \"076cd1d5-0f9d-4497-9abf-de7dadd41ca9\") " Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.151206 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-utilities" (OuterVolumeSpecName: "utilities") pod "076cd1d5-0f9d-4497-9abf-de7dadd41ca9" (UID: "076cd1d5-0f9d-4497-9abf-de7dadd41ca9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.155832 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-kube-api-access-vrjp6" (OuterVolumeSpecName: "kube-api-access-vrjp6") pod "076cd1d5-0f9d-4497-9abf-de7dadd41ca9" (UID: "076cd1d5-0f9d-4497-9abf-de7dadd41ca9"). InnerVolumeSpecName "kube-api-access-vrjp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.197762 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "076cd1d5-0f9d-4497-9abf-de7dadd41ca9" (UID: "076cd1d5-0f9d-4497-9abf-de7dadd41ca9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.253088 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.253137 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.253167 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrjp6\" (UniqueName: \"kubernetes.io/projected/076cd1d5-0f9d-4497-9abf-de7dadd41ca9-kube-api-access-vrjp6\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.507272 4935 generic.go:334] "Generic (PLEG): container finished" podID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerID="29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206" exitCode=0 Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.507337 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wjl2" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.507382 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wjl2" event={"ID":"076cd1d5-0f9d-4497-9abf-de7dadd41ca9","Type":"ContainerDied","Data":"29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206"} Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.507458 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wjl2" event={"ID":"076cd1d5-0f9d-4497-9abf-de7dadd41ca9","Type":"ContainerDied","Data":"dd9cd45db54ba0b3b4c7776048d0b9825d5a770a375d50d64f83fbc38010969b"} Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.507494 4935 scope.go:117] "RemoveContainer" containerID="29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.531956 4935 scope.go:117] "RemoveContainer" containerID="eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.573137 4935 scope.go:117] "RemoveContainer" containerID="3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.575189 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wjl2"] Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.587777 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9wjl2"] Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.614565 4935 scope.go:117] "RemoveContainer" containerID="29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206" Oct 05 09:00:53 crc kubenswrapper[4935]: E1005 09:00:53.615044 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206\": container with ID starting with 29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206 not found: ID does not exist" containerID="29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.615100 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206"} err="failed to get container status \"29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206\": rpc error: code = NotFound desc = could not find container \"29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206\": container with ID starting with 29fd3aad532cfcf9c8e8ed4596b0e4b57242fc727526829dc6e4349a4b011206 not found: ID does not exist" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.615133 4935 scope.go:117] "RemoveContainer" containerID="eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838" Oct 05 09:00:53 crc kubenswrapper[4935]: E1005 09:00:53.615529 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838\": container with ID starting with eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838 not found: ID does not exist" containerID="eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.615578 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838"} err="failed to get container status \"eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838\": rpc error: code = NotFound desc = could not find container \"eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838\": container with ID starting with eeb21222fdc3267f655bddac879a2e8e01aa02cf55559d27d6afc1640dd04838 not found: ID does not exist" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.615627 4935 scope.go:117] "RemoveContainer" containerID="3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423" Oct 05 09:00:53 crc kubenswrapper[4935]: E1005 09:00:53.616094 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423\": container with ID starting with 3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423 not found: ID does not exist" containerID="3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.616134 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423"} err="failed to get container status \"3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423\": rpc error: code = NotFound desc = could not find container \"3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423\": container with ID starting with 3f32dc3356e76b11530fef81fc033b5e57ffe94514b7eea11c03af3fd7edd423 not found: ID does not exist" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.761022 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-x2xbd"] Oct 05 09:00:53 crc kubenswrapper[4935]: E1005 09:00:53.761460 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerName="extract-utilities" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.761482 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerName="extract-utilities" Oct 05 09:00:53 crc kubenswrapper[4935]: E1005 09:00:53.761507 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6dceca-640d-4af2-845d-48741604e927" containerName="mariadb-account-create" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.761515 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6dceca-640d-4af2-845d-48741604e927" containerName="mariadb-account-create" Oct 05 09:00:53 crc kubenswrapper[4935]: E1005 09:00:53.761538 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerName="registry-server" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.761547 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerName="registry-server" Oct 05 09:00:53 crc kubenswrapper[4935]: E1005 09:00:53.761567 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerName="extract-content" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.761574 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerName="extract-content" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.761765 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b6dceca-640d-4af2-845d-48741604e927" containerName="mariadb-account-create" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.761796 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" containerName="registry-server" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.762615 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.768486 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.768670 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-xk5t2" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.768807 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.773510 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-x2xbd"] Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.862978 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v7x8\" (UniqueName: \"kubernetes.io/projected/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-kube-api-access-7v7x8\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.863033 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-combined-ca-bundle\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.863134 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-config-data\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.863151 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-scripts\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.964859 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-config-data\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.964932 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-scripts\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.965085 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v7x8\" (UniqueName: \"kubernetes.io/projected/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-kube-api-access-7v7x8\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.965118 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-combined-ca-bundle\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.970351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-scripts\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.970777 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-combined-ca-bundle\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.971128 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-config-data\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:53 crc kubenswrapper[4935]: I1005 09:00:53.989062 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v7x8\" (UniqueName: \"kubernetes.io/projected/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-kube-api-access-7v7x8\") pod \"aodh-db-sync-x2xbd\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:54 crc kubenswrapper[4935]: I1005 09:00:54.115510 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:00:54 crc kubenswrapper[4935]: I1005 09:00:54.584808 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-x2xbd"] Oct 05 09:00:54 crc kubenswrapper[4935]: W1005 09:00:54.586608 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e0fa5c4_7fee_4b93_8ab1_3eb46da82cc6.slice/crio-e037f6add2383a47f6345c694355c2c72a189e09b55ed5ace6f4f5683415e694 WatchSource:0}: Error finding container e037f6add2383a47f6345c694355c2c72a189e09b55ed5ace6f4f5683415e694: Status 404 returned error can't find the container with id e037f6add2383a47f6345c694355c2c72a189e09b55ed5ace6f4f5683415e694 Oct 05 09:00:54 crc kubenswrapper[4935]: I1005 09:00:54.795418 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="076cd1d5-0f9d-4497-9abf-de7dadd41ca9" path="/var/lib/kubelet/pods/076cd1d5-0f9d-4497-9abf-de7dadd41ca9/volumes" Oct 05 09:00:55 crc kubenswrapper[4935]: I1005 09:00:55.529390 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-x2xbd" event={"ID":"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6","Type":"ContainerStarted","Data":"e037f6add2383a47f6345c694355c2c72a189e09b55ed5ace6f4f5683415e694"} Oct 05 09:00:55 crc kubenswrapper[4935]: I1005 09:00:55.608372 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 09:00:57 crc kubenswrapper[4935]: I1005 09:00:57.777677 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:00:57 crc kubenswrapper[4935]: E1005 09:00:57.778557 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:00:58 crc kubenswrapper[4935]: I1005 09:00:58.036081 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-xlvg7"] Oct 05 09:00:58 crc kubenswrapper[4935]: I1005 09:00:58.043250 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-xlvg7"] Oct 05 09:00:58 crc kubenswrapper[4935]: I1005 09:00:58.792628 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52eec721-ef6b-46f8-a750-287f89c552eb" path="/var/lib/kubelet/pods/52eec721-ef6b-46f8-a750-287f89c552eb/volumes" Oct 05 09:00:59 crc kubenswrapper[4935]: I1005 09:00:59.596144 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-x2xbd" event={"ID":"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6","Type":"ContainerStarted","Data":"3f6ff4ea375bafaa059943b0faae2faee9d9728eb79375bda1e812f961bce74b"} Oct 05 09:00:59 crc kubenswrapper[4935]: I1005 09:00:59.615546 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-x2xbd" podStartSLOduration=2.095757692 podStartE2EDuration="6.615528372s" podCreationTimestamp="2025-10-05 09:00:53 +0000 UTC" firstStartedPulling="2025-10-05 09:00:54.590186268 +0000 UTC m=+7688.472812728" lastFinishedPulling="2025-10-05 09:00:59.109956948 +0000 UTC m=+7692.992583408" observedRunningTime="2025-10-05 09:00:59.614546486 +0000 UTC m=+7693.497172956" watchObservedRunningTime="2025-10-05 09:00:59.615528372 +0000 UTC m=+7693.498154842" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.145320 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29327581-lgtqk"] Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.147038 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.160814 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29327581-lgtqk"] Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.301774 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-combined-ca-bundle\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.302234 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7rhn\" (UniqueName: \"kubernetes.io/projected/57010d7c-9499-4337-b1de-29766c39327a-kube-api-access-g7rhn\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.302496 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-config-data\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.302687 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-fernet-keys\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.405216 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-config-data\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.405374 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-fernet-keys\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.405556 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-combined-ca-bundle\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.405786 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7rhn\" (UniqueName: \"kubernetes.io/projected/57010d7c-9499-4337-b1de-29766c39327a-kube-api-access-g7rhn\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.428461 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-fernet-keys\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.428868 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-config-data\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.429701 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-combined-ca-bundle\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.441641 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7rhn\" (UniqueName: \"kubernetes.io/projected/57010d7c-9499-4337-b1de-29766c39327a-kube-api-access-g7rhn\") pod \"keystone-cron-29327581-lgtqk\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.469769 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:00 crc kubenswrapper[4935]: I1005 09:01:00.984226 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29327581-lgtqk"] Oct 05 09:01:00 crc kubenswrapper[4935]: W1005 09:01:00.991876 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57010d7c_9499_4337_b1de_29766c39327a.slice/crio-88cdf27e6a0cbb784a212dac60a7a798da716376414072a25976685d87318023 WatchSource:0}: Error finding container 88cdf27e6a0cbb784a212dac60a7a798da716376414072a25976685d87318023: Status 404 returned error can't find the container with id 88cdf27e6a0cbb784a212dac60a7a798da716376414072a25976685d87318023 Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.008130 4935 scope.go:117] "RemoveContainer" containerID="2a3c4d5d5b222b0fd1149801e3d85fa52afa7225192c25dabe55b80f02545549" Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.087036 4935 scope.go:117] "RemoveContainer" containerID="3e9e1b7c7535382a5b8bbdee4c249266afa59ca2ab5beb23bd1c8446bc01cc9e" Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.133657 4935 scope.go:117] "RemoveContainer" containerID="7207f2634bb3213d44cec07ddb4f76cf1d2050277444e2b079c788b06155dc5f" Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.161875 4935 scope.go:117] "RemoveContainer" containerID="42a1cf7e66fb4fe2e77524328722fcb1608733769fcbcd0f661d15c22748c0ad" Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.187189 4935 scope.go:117] "RemoveContainer" containerID="a93fa822736fd8b2807064fd05ce98c2aae5ae3d7cf223827499f4add0414a4e" Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.619764 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327581-lgtqk" event={"ID":"57010d7c-9499-4337-b1de-29766c39327a","Type":"ContainerStarted","Data":"c37b5de21caf00be54144f9ac78319edeb041f186b27bf3a4177877768243aa3"} Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.620102 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327581-lgtqk" event={"ID":"57010d7c-9499-4337-b1de-29766c39327a","Type":"ContainerStarted","Data":"88cdf27e6a0cbb784a212dac60a7a798da716376414072a25976685d87318023"} Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.622172 4935 generic.go:334] "Generic (PLEG): container finished" podID="6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" containerID="3f6ff4ea375bafaa059943b0faae2faee9d9728eb79375bda1e812f961bce74b" exitCode=0 Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.622222 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-x2xbd" event={"ID":"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6","Type":"ContainerDied","Data":"3f6ff4ea375bafaa059943b0faae2faee9d9728eb79375bda1e812f961bce74b"} Oct 05 09:01:01 crc kubenswrapper[4935]: I1005 09:01:01.648955 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29327581-lgtqk" podStartSLOduration=1.648934624 podStartE2EDuration="1.648934624s" podCreationTimestamp="2025-10-05 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:01:01.64273653 +0000 UTC m=+7695.525362990" watchObservedRunningTime="2025-10-05 09:01:01.648934624 +0000 UTC m=+7695.531561104" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.121615 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.174622 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-scripts\") pod \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.174767 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-config-data\") pod \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.174841 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-combined-ca-bundle\") pod \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.174997 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v7x8\" (UniqueName: \"kubernetes.io/projected/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-kube-api-access-7v7x8\") pod \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\" (UID: \"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6\") " Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.183189 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-kube-api-access-7v7x8" (OuterVolumeSpecName: "kube-api-access-7v7x8") pod "6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" (UID: "6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6"). InnerVolumeSpecName "kube-api-access-7v7x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.187013 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-scripts" (OuterVolumeSpecName: "scripts") pod "6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" (UID: "6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.209916 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" (UID: "6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.212384 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-config-data" (OuterVolumeSpecName: "config-data") pod "6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" (UID: "6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.279190 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.279230 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.279246 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v7x8\" (UniqueName: \"kubernetes.io/projected/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-kube-api-access-7v7x8\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.279259 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.652931 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-x2xbd" event={"ID":"6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6","Type":"ContainerDied","Data":"e037f6add2383a47f6345c694355c2c72a189e09b55ed5ace6f4f5683415e694"} Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.652981 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e037f6add2383a47f6345c694355c2c72a189e09b55ed5ace6f4f5683415e694" Oct 05 09:01:03 crc kubenswrapper[4935]: I1005 09:01:03.653035 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-x2xbd" Oct 05 09:01:04 crc kubenswrapper[4935]: I1005 09:01:04.670538 4935 generic.go:334] "Generic (PLEG): container finished" podID="57010d7c-9499-4337-b1de-29766c39327a" containerID="c37b5de21caf00be54144f9ac78319edeb041f186b27bf3a4177877768243aa3" exitCode=0 Oct 05 09:01:04 crc kubenswrapper[4935]: I1005 09:01:04.670624 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327581-lgtqk" event={"ID":"57010d7c-9499-4337-b1de-29766c39327a","Type":"ContainerDied","Data":"c37b5de21caf00be54144f9ac78319edeb041f186b27bf3a4177877768243aa3"} Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.142017 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.271611 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-fernet-keys\") pod \"57010d7c-9499-4337-b1de-29766c39327a\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.271694 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-combined-ca-bundle\") pod \"57010d7c-9499-4337-b1de-29766c39327a\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.271723 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7rhn\" (UniqueName: \"kubernetes.io/projected/57010d7c-9499-4337-b1de-29766c39327a-kube-api-access-g7rhn\") pod \"57010d7c-9499-4337-b1de-29766c39327a\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.271747 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-config-data\") pod \"57010d7c-9499-4337-b1de-29766c39327a\" (UID: \"57010d7c-9499-4337-b1de-29766c39327a\") " Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.279035 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57010d7c-9499-4337-b1de-29766c39327a-kube-api-access-g7rhn" (OuterVolumeSpecName: "kube-api-access-g7rhn") pod "57010d7c-9499-4337-b1de-29766c39327a" (UID: "57010d7c-9499-4337-b1de-29766c39327a"). InnerVolumeSpecName "kube-api-access-g7rhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.279400 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "57010d7c-9499-4337-b1de-29766c39327a" (UID: "57010d7c-9499-4337-b1de-29766c39327a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.322324 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57010d7c-9499-4337-b1de-29766c39327a" (UID: "57010d7c-9499-4337-b1de-29766c39327a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.345019 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-config-data" (OuterVolumeSpecName: "config-data") pod "57010d7c-9499-4337-b1de-29766c39327a" (UID: "57010d7c-9499-4337-b1de-29766c39327a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.373983 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.374020 4935 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.374035 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57010d7c-9499-4337-b1de-29766c39327a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.374050 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7rhn\" (UniqueName: \"kubernetes.io/projected/57010d7c-9499-4337-b1de-29766c39327a-kube-api-access-g7rhn\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.698627 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327581-lgtqk" event={"ID":"57010d7c-9499-4337-b1de-29766c39327a","Type":"ContainerDied","Data":"88cdf27e6a0cbb784a212dac60a7a798da716376414072a25976685d87318023"} Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.698687 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88cdf27e6a0cbb784a212dac60a7a798da716376414072a25976685d87318023" Oct 05 09:01:06 crc kubenswrapper[4935]: I1005 09:01:06.698770 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327581-lgtqk" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.325403 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 05 09:01:08 crc kubenswrapper[4935]: E1005 09:01:08.326244 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57010d7c-9499-4337-b1de-29766c39327a" containerName="keystone-cron" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.326260 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="57010d7c-9499-4337-b1de-29766c39327a" containerName="keystone-cron" Oct 05 09:01:08 crc kubenswrapper[4935]: E1005 09:01:08.326272 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" containerName="aodh-db-sync" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.326278 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" containerName="aodh-db-sync" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.326503 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" containerName="aodh-db-sync" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.326530 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="57010d7c-9499-4337-b1de-29766c39327a" containerName="keystone-cron" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.328465 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.331613 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-xk5t2" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.331690 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.331821 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.351437 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.416578 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.416703 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-scripts\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.416766 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dw98\" (UniqueName: \"kubernetes.io/projected/b6e988f2-3f23-4152-9830-5e03d950de38-kube-api-access-5dw98\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.417353 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-config-data\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.519405 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-config-data\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.519843 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.519948 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-scripts\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.519986 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dw98\" (UniqueName: \"kubernetes.io/projected/b6e988f2-3f23-4152-9830-5e03d950de38-kube-api-access-5dw98\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.525412 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.534089 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-scripts\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.535052 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e988f2-3f23-4152-9830-5e03d950de38-config-data\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.545391 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dw98\" (UniqueName: \"kubernetes.io/projected/b6e988f2-3f23-4152-9830-5e03d950de38-kube-api-access-5dw98\") pod \"aodh-0\" (UID: \"b6e988f2-3f23-4152-9830-5e03d950de38\") " pod="openstack/aodh-0" Oct 05 09:01:08 crc kubenswrapper[4935]: I1005 09:01:08.672474 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 05 09:01:09 crc kubenswrapper[4935]: I1005 09:01:09.286967 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 05 09:01:09 crc kubenswrapper[4935]: I1005 09:01:09.744010 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b6e988f2-3f23-4152-9830-5e03d950de38","Type":"ContainerStarted","Data":"dc53b63a12bc4152dc3b590aa86973703ee3dfe49a9bdc4972dbedc9d27e59bf"} Oct 05 09:01:10 crc kubenswrapper[4935]: I1005 09:01:10.763884 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b6e988f2-3f23-4152-9830-5e03d950de38","Type":"ContainerStarted","Data":"0750027584f1a7eabc7d505ea108f1916e55f19ffbb66853b0527a3a0c21581d"} Oct 05 09:01:10 crc kubenswrapper[4935]: I1005 09:01:10.893133 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:01:10 crc kubenswrapper[4935]: I1005 09:01:10.893410 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="ceilometer-central-agent" containerID="cri-o://b0f379a800d47443b8e9e414ae37a0d586a6e0469fd7b7ff86350d154bb6bbb7" gracePeriod=30 Oct 05 09:01:10 crc kubenswrapper[4935]: I1005 09:01:10.895680 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="sg-core" containerID="cri-o://32c2df667e2bf45da5733e1a0fe2831cc1607a2b6b71f7b0b75a3d97463fd1bc" gracePeriod=30 Oct 05 09:01:10 crc kubenswrapper[4935]: I1005 09:01:10.896019 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="proxy-httpd" containerID="cri-o://9a10344778c84fa6aee5f7c67d6fefdf14ffe8c59ff9c56e68d7e69dfb61452b" gracePeriod=30 Oct 05 09:01:10 crc kubenswrapper[4935]: I1005 09:01:10.896034 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="ceilometer-notification-agent" containerID="cri-o://8be51b3e4f763e7715afeb82eee38445aaca2dc6740b1be143f7d4be727101aa" gracePeriod=30 Oct 05 09:01:11 crc kubenswrapper[4935]: I1005 09:01:11.781717 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b6e988f2-3f23-4152-9830-5e03d950de38","Type":"ContainerStarted","Data":"8a502644c137041bea69103f2928c2e94fa9571af6adc6cf2ddd0929563143b4"} Oct 05 09:01:11 crc kubenswrapper[4935]: I1005 09:01:11.781742 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:01:11 crc kubenswrapper[4935]: E1005 09:01:11.783091 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:01:11 crc kubenswrapper[4935]: I1005 09:01:11.790526 4935 generic.go:334] "Generic (PLEG): container finished" podID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerID="9a10344778c84fa6aee5f7c67d6fefdf14ffe8c59ff9c56e68d7e69dfb61452b" exitCode=0 Oct 05 09:01:11 crc kubenswrapper[4935]: I1005 09:01:11.790566 4935 generic.go:334] "Generic (PLEG): container finished" podID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerID="32c2df667e2bf45da5733e1a0fe2831cc1607a2b6b71f7b0b75a3d97463fd1bc" exitCode=2 Oct 05 09:01:11 crc kubenswrapper[4935]: I1005 09:01:11.790578 4935 generic.go:334] "Generic (PLEG): container finished" podID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerID="b0f379a800d47443b8e9e414ae37a0d586a6e0469fd7b7ff86350d154bb6bbb7" exitCode=0 Oct 05 09:01:11 crc kubenswrapper[4935]: I1005 09:01:11.790603 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerDied","Data":"9a10344778c84fa6aee5f7c67d6fefdf14ffe8c59ff9c56e68d7e69dfb61452b"} Oct 05 09:01:11 crc kubenswrapper[4935]: I1005 09:01:11.790676 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerDied","Data":"32c2df667e2bf45da5733e1a0fe2831cc1607a2b6b71f7b0b75a3d97463fd1bc"} Oct 05 09:01:11 crc kubenswrapper[4935]: I1005 09:01:11.790694 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerDied","Data":"b0f379a800d47443b8e9e414ae37a0d586a6e0469fd7b7ff86350d154bb6bbb7"} Oct 05 09:01:12 crc kubenswrapper[4935]: I1005 09:01:12.814183 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b6e988f2-3f23-4152-9830-5e03d950de38","Type":"ContainerStarted","Data":"38bdeb03614c6ab4f16237fdc0557611867ed809b97273ce52eed97071106218"} Oct 05 09:01:13 crc kubenswrapper[4935]: I1005 09:01:13.836858 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b6e988f2-3f23-4152-9830-5e03d950de38","Type":"ContainerStarted","Data":"027b8d52cf71e6caddd8fb66a3deefe297b363056245a650718465d85b78be84"} Oct 05 09:01:13 crc kubenswrapper[4935]: I1005 09:01:13.867699 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.7589080400000001 podStartE2EDuration="5.867675364s" podCreationTimestamp="2025-10-05 09:01:08 +0000 UTC" firstStartedPulling="2025-10-05 09:01:09.296232173 +0000 UTC m=+7703.178858633" lastFinishedPulling="2025-10-05 09:01:13.404999497 +0000 UTC m=+7707.287625957" observedRunningTime="2025-10-05 09:01:13.867059688 +0000 UTC m=+7707.749686148" watchObservedRunningTime="2025-10-05 09:01:13.867675364 +0000 UTC m=+7707.750301824" Oct 05 09:01:14 crc kubenswrapper[4935]: I1005 09:01:14.869558 4935 generic.go:334] "Generic (PLEG): container finished" podID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerID="8be51b3e4f763e7715afeb82eee38445aaca2dc6740b1be143f7d4be727101aa" exitCode=0 Oct 05 09:01:14 crc kubenswrapper[4935]: I1005 09:01:14.869666 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerDied","Data":"8be51b3e4f763e7715afeb82eee38445aaca2dc6740b1be143f7d4be727101aa"} Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.090748 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.180394 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-config-data\") pod \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.180531 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-sg-core-conf-yaml\") pod \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.180561 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp2hd\" (UniqueName: \"kubernetes.io/projected/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-kube-api-access-vp2hd\") pod \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.180612 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-scripts\") pod \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.180639 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-run-httpd\") pod \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.180778 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-log-httpd\") pod \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.181014 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-combined-ca-bundle\") pod \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\" (UID: \"2f3e3128-ff4a-4657-ba2b-643069a0fc8c\") " Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.181856 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2f3e3128-ff4a-4657-ba2b-643069a0fc8c" (UID: "2f3e3128-ff4a-4657-ba2b-643069a0fc8c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.182054 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2f3e3128-ff4a-4657-ba2b-643069a0fc8c" (UID: "2f3e3128-ff4a-4657-ba2b-643069a0fc8c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.182161 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.182174 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.190032 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-scripts" (OuterVolumeSpecName: "scripts") pod "2f3e3128-ff4a-4657-ba2b-643069a0fc8c" (UID: "2f3e3128-ff4a-4657-ba2b-643069a0fc8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.204594 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-kube-api-access-vp2hd" (OuterVolumeSpecName: "kube-api-access-vp2hd") pod "2f3e3128-ff4a-4657-ba2b-643069a0fc8c" (UID: "2f3e3128-ff4a-4657-ba2b-643069a0fc8c"). InnerVolumeSpecName "kube-api-access-vp2hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.216111 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2f3e3128-ff4a-4657-ba2b-643069a0fc8c" (UID: "2f3e3128-ff4a-4657-ba2b-643069a0fc8c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.254445 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f3e3128-ff4a-4657-ba2b-643069a0fc8c" (UID: "2f3e3128-ff4a-4657-ba2b-643069a0fc8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.284064 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.284112 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp2hd\" (UniqueName: \"kubernetes.io/projected/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-kube-api-access-vp2hd\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.284124 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.284136 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.291069 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-config-data" (OuterVolumeSpecName: "config-data") pod "2f3e3128-ff4a-4657-ba2b-643069a0fc8c" (UID: "2f3e3128-ff4a-4657-ba2b-643069a0fc8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.386535 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f3e3128-ff4a-4657-ba2b-643069a0fc8c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.881068 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f3e3128-ff4a-4657-ba2b-643069a0fc8c","Type":"ContainerDied","Data":"9069cc1d9e934e88037fbaf3b8c7b9eeb6151755dfb88a46f123ce742d0ea3e2"} Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.881123 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.881135 4935 scope.go:117] "RemoveContainer" containerID="9a10344778c84fa6aee5f7c67d6fefdf14ffe8c59ff9c56e68d7e69dfb61452b" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.919255 4935 scope.go:117] "RemoveContainer" containerID="32c2df667e2bf45da5733e1a0fe2831cc1607a2b6b71f7b0b75a3d97463fd1bc" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.940585 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.950911 4935 scope.go:117] "RemoveContainer" containerID="8be51b3e4f763e7715afeb82eee38445aaca2dc6740b1be143f7d4be727101aa" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.952464 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.965987 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:01:15 crc kubenswrapper[4935]: E1005 09:01:15.966438 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="ceilometer-notification-agent" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.966455 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="ceilometer-notification-agent" Oct 05 09:01:15 crc kubenswrapper[4935]: E1005 09:01:15.966478 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="ceilometer-central-agent" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.966485 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="ceilometer-central-agent" Oct 05 09:01:15 crc kubenswrapper[4935]: E1005 09:01:15.966497 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="sg-core" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.966504 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="sg-core" Oct 05 09:01:15 crc kubenswrapper[4935]: E1005 09:01:15.966530 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="proxy-httpd" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.966538 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="proxy-httpd" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.966758 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="ceilometer-notification-agent" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.966786 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="sg-core" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.966801 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="ceilometer-central-agent" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.966817 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" containerName="proxy-httpd" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.968778 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.973560 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.973676 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.977118 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:01:15 crc kubenswrapper[4935]: I1005 09:01:15.998950 4935 scope.go:117] "RemoveContainer" containerID="b0f379a800d47443b8e9e414ae37a0d586a6e0469fd7b7ff86350d154bb6bbb7" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.106065 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.106131 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-log-httpd\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.106206 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-config-data\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.106235 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.106268 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-run-httpd\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.106298 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-scripts\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.106328 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d822\" (UniqueName: \"kubernetes.io/projected/0b167df5-0f56-489f-bb85-d48128d3ba07-kube-api-access-7d822\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.208264 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-log-httpd\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.208682 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-log-httpd\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.208843 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-config-data\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.208876 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.208958 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-run-httpd\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.209516 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-run-httpd\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.209599 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-scripts\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.209641 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d822\" (UniqueName: \"kubernetes.io/projected/0b167df5-0f56-489f-bb85-d48128d3ba07-kube-api-access-7d822\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.210040 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.213148 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.218114 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.219627 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-scripts\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.228649 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d822\" (UniqueName: \"kubernetes.io/projected/0b167df5-0f56-489f-bb85-d48128d3ba07-kube-api-access-7d822\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.228990 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-config-data\") pod \"ceilometer-0\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.286834 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.765773 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.794092 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f3e3128-ff4a-4657-ba2b-643069a0fc8c" path="/var/lib/kubelet/pods/2f3e3128-ff4a-4657-ba2b-643069a0fc8c/volumes" Oct 05 09:01:16 crc kubenswrapper[4935]: I1005 09:01:16.890914 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerStarted","Data":"b9a9e6ffad92a603742435ad66a5cb131ea39e4abf9a5cad03d030b0f531d798"} Oct 05 09:01:17 crc kubenswrapper[4935]: I1005 09:01:17.905591 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerStarted","Data":"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761"} Oct 05 09:01:17 crc kubenswrapper[4935]: I1005 09:01:17.907089 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerStarted","Data":"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111"} Oct 05 09:01:18 crc kubenswrapper[4935]: I1005 09:01:18.917400 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerStarted","Data":"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb"} Oct 05 09:01:18 crc kubenswrapper[4935]: I1005 09:01:18.931209 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-kncpn"] Oct 05 09:01:18 crc kubenswrapper[4935]: I1005 09:01:18.933237 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-kncpn" Oct 05 09:01:18 crc kubenswrapper[4935]: I1005 09:01:18.940001 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-kncpn"] Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.075662 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bx4s\" (UniqueName: \"kubernetes.io/projected/72a208bb-52b8-43e3-b8f4-21511674b129-kube-api-access-9bx4s\") pod \"manila-db-create-kncpn\" (UID: \"72a208bb-52b8-43e3-b8f4-21511674b129\") " pod="openstack/manila-db-create-kncpn" Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.178145 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bx4s\" (UniqueName: \"kubernetes.io/projected/72a208bb-52b8-43e3-b8f4-21511674b129-kube-api-access-9bx4s\") pod \"manila-db-create-kncpn\" (UID: \"72a208bb-52b8-43e3-b8f4-21511674b129\") " pod="openstack/manila-db-create-kncpn" Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.201355 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bx4s\" (UniqueName: \"kubernetes.io/projected/72a208bb-52b8-43e3-b8f4-21511674b129-kube-api-access-9bx4s\") pod \"manila-db-create-kncpn\" (UID: \"72a208bb-52b8-43e3-b8f4-21511674b129\") " pod="openstack/manila-db-create-kncpn" Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.273927 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-kncpn" Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.832484 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-kncpn"] Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.933370 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-kncpn" event={"ID":"72a208bb-52b8-43e3-b8f4-21511674b129","Type":"ContainerStarted","Data":"0dc6526d64ff71cd058ba9e4dece89af5bcec3bcdfe6f8cc2db63a786f5e691a"} Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.936482 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerStarted","Data":"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859"} Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.936693 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 09:01:19 crc kubenswrapper[4935]: I1005 09:01:19.964526 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.205268202 podStartE2EDuration="4.964504217s" podCreationTimestamp="2025-10-05 09:01:15 +0000 UTC" firstStartedPulling="2025-10-05 09:01:16.776770552 +0000 UTC m=+7710.659397012" lastFinishedPulling="2025-10-05 09:01:19.536006557 +0000 UTC m=+7713.418633027" observedRunningTime="2025-10-05 09:01:19.95671418 +0000 UTC m=+7713.839340640" watchObservedRunningTime="2025-10-05 09:01:19.964504217 +0000 UTC m=+7713.847130677" Oct 05 09:01:20 crc kubenswrapper[4935]: I1005 09:01:20.947439 4935 generic.go:334] "Generic (PLEG): container finished" podID="72a208bb-52b8-43e3-b8f4-21511674b129" containerID="220670d55e5cfad5d7b830c25e329103c382da0fb77952d4cec10e7549a4b9ea" exitCode=0 Oct 05 09:01:20 crc kubenswrapper[4935]: I1005 09:01:20.947551 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-kncpn" event={"ID":"72a208bb-52b8-43e3-b8f4-21511674b129","Type":"ContainerDied","Data":"220670d55e5cfad5d7b830c25e329103c382da0fb77952d4cec10e7549a4b9ea"} Oct 05 09:01:22 crc kubenswrapper[4935]: I1005 09:01:22.534278 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-kncpn" Oct 05 09:01:22 crc kubenswrapper[4935]: I1005 09:01:22.662679 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bx4s\" (UniqueName: \"kubernetes.io/projected/72a208bb-52b8-43e3-b8f4-21511674b129-kube-api-access-9bx4s\") pod \"72a208bb-52b8-43e3-b8f4-21511674b129\" (UID: \"72a208bb-52b8-43e3-b8f4-21511674b129\") " Oct 05 09:01:22 crc kubenswrapper[4935]: I1005 09:01:22.675127 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72a208bb-52b8-43e3-b8f4-21511674b129-kube-api-access-9bx4s" (OuterVolumeSpecName: "kube-api-access-9bx4s") pod "72a208bb-52b8-43e3-b8f4-21511674b129" (UID: "72a208bb-52b8-43e3-b8f4-21511674b129"). InnerVolumeSpecName "kube-api-access-9bx4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:22 crc kubenswrapper[4935]: I1005 09:01:22.764961 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bx4s\" (UniqueName: \"kubernetes.io/projected/72a208bb-52b8-43e3-b8f4-21511674b129-kube-api-access-9bx4s\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:22 crc kubenswrapper[4935]: I1005 09:01:22.971347 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-kncpn" event={"ID":"72a208bb-52b8-43e3-b8f4-21511674b129","Type":"ContainerDied","Data":"0dc6526d64ff71cd058ba9e4dece89af5bcec3bcdfe6f8cc2db63a786f5e691a"} Oct 05 09:01:22 crc kubenswrapper[4935]: I1005 09:01:22.971397 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dc6526d64ff71cd058ba9e4dece89af5bcec3bcdfe6f8cc2db63a786f5e691a" Oct 05 09:01:22 crc kubenswrapper[4935]: I1005 09:01:22.971421 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-kncpn" Oct 05 09:01:24 crc kubenswrapper[4935]: I1005 09:01:24.776972 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:01:24 crc kubenswrapper[4935]: E1005 09:01:24.777489 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.058819 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-00a5-account-create-f7wmx"] Oct 05 09:01:29 crc kubenswrapper[4935]: E1005 09:01:29.060219 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a208bb-52b8-43e3-b8f4-21511674b129" containerName="mariadb-database-create" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.060242 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a208bb-52b8-43e3-b8f4-21511674b129" containerName="mariadb-database-create" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.060602 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a208bb-52b8-43e3-b8f4-21511674b129" containerName="mariadb-database-create" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.062053 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-00a5-account-create-f7wmx" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.068121 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-00a5-account-create-f7wmx"] Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.097659 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.200760 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg55t\" (UniqueName: \"kubernetes.io/projected/6debd557-00e3-424d-b70a-f94f1651f8f0-kube-api-access-bg55t\") pod \"manila-00a5-account-create-f7wmx\" (UID: \"6debd557-00e3-424d-b70a-f94f1651f8f0\") " pod="openstack/manila-00a5-account-create-f7wmx" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.302992 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg55t\" (UniqueName: \"kubernetes.io/projected/6debd557-00e3-424d-b70a-f94f1651f8f0-kube-api-access-bg55t\") pod \"manila-00a5-account-create-f7wmx\" (UID: \"6debd557-00e3-424d-b70a-f94f1651f8f0\") " pod="openstack/manila-00a5-account-create-f7wmx" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.333979 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg55t\" (UniqueName: \"kubernetes.io/projected/6debd557-00e3-424d-b70a-f94f1651f8f0-kube-api-access-bg55t\") pod \"manila-00a5-account-create-f7wmx\" (UID: \"6debd557-00e3-424d-b70a-f94f1651f8f0\") " pod="openstack/manila-00a5-account-create-f7wmx" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.422591 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-00a5-account-create-f7wmx" Oct 05 09:01:29 crc kubenswrapper[4935]: I1005 09:01:29.931411 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-00a5-account-create-f7wmx"] Oct 05 09:01:29 crc kubenswrapper[4935]: W1005 09:01:29.934130 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6debd557_00e3_424d_b70a_f94f1651f8f0.slice/crio-13f79ccf828574a126a2a7ab933f9c625d3a433b57853ef86296e2c871928049 WatchSource:0}: Error finding container 13f79ccf828574a126a2a7ab933f9c625d3a433b57853ef86296e2c871928049: Status 404 returned error can't find the container with id 13f79ccf828574a126a2a7ab933f9c625d3a433b57853ef86296e2c871928049 Oct 05 09:01:30 crc kubenswrapper[4935]: I1005 09:01:30.112940 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-00a5-account-create-f7wmx" event={"ID":"6debd557-00e3-424d-b70a-f94f1651f8f0","Type":"ContainerStarted","Data":"13f79ccf828574a126a2a7ab933f9c625d3a433b57853ef86296e2c871928049"} Oct 05 09:01:31 crc kubenswrapper[4935]: I1005 09:01:31.125433 4935 generic.go:334] "Generic (PLEG): container finished" podID="6debd557-00e3-424d-b70a-f94f1651f8f0" containerID="d60eeb5c910f425f5c1afe5c3640ca78036735a2a16d87ec97b6cb4cf77394b2" exitCode=0 Oct 05 09:01:31 crc kubenswrapper[4935]: I1005 09:01:31.125514 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-00a5-account-create-f7wmx" event={"ID":"6debd557-00e3-424d-b70a-f94f1651f8f0","Type":"ContainerDied","Data":"d60eeb5c910f425f5c1afe5c3640ca78036735a2a16d87ec97b6cb4cf77394b2"} Oct 05 09:01:32 crc kubenswrapper[4935]: I1005 09:01:32.570014 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-00a5-account-create-f7wmx" Oct 05 09:01:32 crc kubenswrapper[4935]: I1005 09:01:32.678426 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg55t\" (UniqueName: \"kubernetes.io/projected/6debd557-00e3-424d-b70a-f94f1651f8f0-kube-api-access-bg55t\") pod \"6debd557-00e3-424d-b70a-f94f1651f8f0\" (UID: \"6debd557-00e3-424d-b70a-f94f1651f8f0\") " Oct 05 09:01:32 crc kubenswrapper[4935]: I1005 09:01:32.697758 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6debd557-00e3-424d-b70a-f94f1651f8f0-kube-api-access-bg55t" (OuterVolumeSpecName: "kube-api-access-bg55t") pod "6debd557-00e3-424d-b70a-f94f1651f8f0" (UID: "6debd557-00e3-424d-b70a-f94f1651f8f0"). InnerVolumeSpecName "kube-api-access-bg55t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:32 crc kubenswrapper[4935]: I1005 09:01:32.781814 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg55t\" (UniqueName: \"kubernetes.io/projected/6debd557-00e3-424d-b70a-f94f1651f8f0-kube-api-access-bg55t\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:33 crc kubenswrapper[4935]: I1005 09:01:33.149283 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-00a5-account-create-f7wmx" event={"ID":"6debd557-00e3-424d-b70a-f94f1651f8f0","Type":"ContainerDied","Data":"13f79ccf828574a126a2a7ab933f9c625d3a433b57853ef86296e2c871928049"} Oct 05 09:01:33 crc kubenswrapper[4935]: I1005 09:01:33.149328 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-00a5-account-create-f7wmx" Oct 05 09:01:33 crc kubenswrapper[4935]: I1005 09:01:33.149342 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13f79ccf828574a126a2a7ab933f9c625d3a433b57853ef86296e2c871928049" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.401271 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-wn8z7"] Oct 05 09:01:34 crc kubenswrapper[4935]: E1005 09:01:34.401986 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6debd557-00e3-424d-b70a-f94f1651f8f0" containerName="mariadb-account-create" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.402000 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6debd557-00e3-424d-b70a-f94f1651f8f0" containerName="mariadb-account-create" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.402239 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6debd557-00e3-424d-b70a-f94f1651f8f0" containerName="mariadb-account-create" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.403068 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.405009 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-4ltpz" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.410762 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.421735 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-wn8z7"] Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.514393 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-combined-ca-bundle\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.514671 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-job-config-data\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.514753 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99xqm\" (UniqueName: \"kubernetes.io/projected/0315ba9d-0e75-4013-b5d0-07728f5d07a0-kube-api-access-99xqm\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.514908 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-config-data\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.616388 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-config-data\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.616485 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-combined-ca-bundle\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.616578 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-job-config-data\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.616727 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99xqm\" (UniqueName: \"kubernetes.io/projected/0315ba9d-0e75-4013-b5d0-07728f5d07a0-kube-api-access-99xqm\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.621331 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-config-data\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.621366 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-combined-ca-bundle\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.627410 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-job-config-data\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.635967 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99xqm\" (UniqueName: \"kubernetes.io/projected/0315ba9d-0e75-4013-b5d0-07728f5d07a0-kube-api-access-99xqm\") pod \"manila-db-sync-wn8z7\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:34 crc kubenswrapper[4935]: I1005 09:01:34.721460 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:35 crc kubenswrapper[4935]: I1005 09:01:35.401505 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-wn8z7"] Oct 05 09:01:36 crc kubenswrapper[4935]: I1005 09:01:36.182861 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-wn8z7" event={"ID":"0315ba9d-0e75-4013-b5d0-07728f5d07a0","Type":"ContainerStarted","Data":"f292627eac5923e95cafe97764d7f31b902e031afd021924b08e435d9f72e2b6"} Oct 05 09:01:39 crc kubenswrapper[4935]: I1005 09:01:39.777024 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:01:39 crc kubenswrapper[4935]: E1005 09:01:39.777855 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:01:41 crc kubenswrapper[4935]: I1005 09:01:41.237713 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-wn8z7" event={"ID":"0315ba9d-0e75-4013-b5d0-07728f5d07a0","Type":"ContainerStarted","Data":"638c462666eed8bd0c52130cadcebd82727a70fddb0eb0ebf09d0d6918c5cd3f"} Oct 05 09:01:41 crc kubenswrapper[4935]: I1005 09:01:41.277355 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-wn8z7" podStartSLOduration=2.31005513 podStartE2EDuration="7.277330316s" podCreationTimestamp="2025-10-05 09:01:34 +0000 UTC" firstStartedPulling="2025-10-05 09:01:35.407738728 +0000 UTC m=+7729.290365188" lastFinishedPulling="2025-10-05 09:01:40.375013874 +0000 UTC m=+7734.257640374" observedRunningTime="2025-10-05 09:01:41.274181892 +0000 UTC m=+7735.156808362" watchObservedRunningTime="2025-10-05 09:01:41.277330316 +0000 UTC m=+7735.159956796" Oct 05 09:01:43 crc kubenswrapper[4935]: I1005 09:01:43.281062 4935 generic.go:334] "Generic (PLEG): container finished" podID="0315ba9d-0e75-4013-b5d0-07728f5d07a0" containerID="638c462666eed8bd0c52130cadcebd82727a70fddb0eb0ebf09d0d6918c5cd3f" exitCode=0 Oct 05 09:01:43 crc kubenswrapper[4935]: I1005 09:01:43.281153 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-wn8z7" event={"ID":"0315ba9d-0e75-4013-b5d0-07728f5d07a0","Type":"ContainerDied","Data":"638c462666eed8bd0c52130cadcebd82727a70fddb0eb0ebf09d0d6918c5cd3f"} Oct 05 09:01:44 crc kubenswrapper[4935]: I1005 09:01:44.815358 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:44 crc kubenswrapper[4935]: I1005 09:01:44.975061 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-combined-ca-bundle\") pod \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " Oct 05 09:01:44 crc kubenswrapper[4935]: I1005 09:01:44.975173 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-job-config-data\") pod \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " Oct 05 09:01:44 crc kubenswrapper[4935]: I1005 09:01:44.975250 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-config-data\") pod \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " Oct 05 09:01:44 crc kubenswrapper[4935]: I1005 09:01:44.975333 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99xqm\" (UniqueName: \"kubernetes.io/projected/0315ba9d-0e75-4013-b5d0-07728f5d07a0-kube-api-access-99xqm\") pod \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\" (UID: \"0315ba9d-0e75-4013-b5d0-07728f5d07a0\") " Oct 05 09:01:44 crc kubenswrapper[4935]: I1005 09:01:44.980744 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0315ba9d-0e75-4013-b5d0-07728f5d07a0-kube-api-access-99xqm" (OuterVolumeSpecName: "kube-api-access-99xqm") pod "0315ba9d-0e75-4013-b5d0-07728f5d07a0" (UID: "0315ba9d-0e75-4013-b5d0-07728f5d07a0"). InnerVolumeSpecName "kube-api-access-99xqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:44 crc kubenswrapper[4935]: I1005 09:01:44.982665 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "0315ba9d-0e75-4013-b5d0-07728f5d07a0" (UID: "0315ba9d-0e75-4013-b5d0-07728f5d07a0"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:44 crc kubenswrapper[4935]: I1005 09:01:44.993748 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-config-data" (OuterVolumeSpecName: "config-data") pod "0315ba9d-0e75-4013-b5d0-07728f5d07a0" (UID: "0315ba9d-0e75-4013-b5d0-07728f5d07a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.006152 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0315ba9d-0e75-4013-b5d0-07728f5d07a0" (UID: "0315ba9d-0e75-4013-b5d0-07728f5d07a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.077561 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.077864 4935 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.077874 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0315ba9d-0e75-4013-b5d0-07728f5d07a0-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.077885 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99xqm\" (UniqueName: \"kubernetes.io/projected/0315ba9d-0e75-4013-b5d0-07728f5d07a0-kube-api-access-99xqm\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.306698 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-wn8z7" event={"ID":"0315ba9d-0e75-4013-b5d0-07728f5d07a0","Type":"ContainerDied","Data":"f292627eac5923e95cafe97764d7f31b902e031afd021924b08e435d9f72e2b6"} Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.306755 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f292627eac5923e95cafe97764d7f31b902e031afd021924b08e435d9f72e2b6" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.306832 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-wn8z7" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.833187 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 05 09:01:45 crc kubenswrapper[4935]: E1005 09:01:45.833925 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0315ba9d-0e75-4013-b5d0-07728f5d07a0" containerName="manila-db-sync" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.833944 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0315ba9d-0e75-4013-b5d0-07728f5d07a0" containerName="manila-db-sync" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.834172 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0315ba9d-0e75-4013-b5d0-07728f5d07a0" containerName="manila-db-sync" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.844545 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.848456 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.849495 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-4ltpz" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.853202 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.853493 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.905506 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.909806 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.917982 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 05 09:01:45 crc kubenswrapper[4935]: I1005 09:01:45.963860 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.012688 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.014119 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4cmr\" (UniqueName: \"kubernetes.io/projected/a3869bdd-ef72-488e-8691-d4c6bc8d8118-kube-api-access-g4cmr\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.014255 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.014338 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.014439 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.015186 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqczh\" (UniqueName: \"kubernetes.io/projected/4340e44d-1160-461e-859a-efebef2f545b-kube-api-access-gqczh\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.015465 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-config-data\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.015669 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3869bdd-ef72-488e-8691-d4c6bc8d8118-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.015920 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-scripts\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.016083 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-scripts\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.016250 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-config-data\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.016501 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a3869bdd-ef72-488e-8691-d4c6bc8d8118-ceph\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.016666 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.016805 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4340e44d-1160-461e-859a-efebef2f545b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.016963 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a3869bdd-ef72-488e-8691-d4c6bc8d8118-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.035132 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77897597d9-bfqsd"] Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.043948 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.065680 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77897597d9-bfqsd"] Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.086187 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.087935 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.089923 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.098655 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.118760 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.118807 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.118851 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.118881 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqczh\" (UniqueName: \"kubernetes.io/projected/4340e44d-1160-461e-859a-efebef2f545b-kube-api-access-gqczh\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.118929 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-config-data\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.118947 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3869bdd-ef72-488e-8691-d4c6bc8d8118-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.118967 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-scripts\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.118984 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-scripts\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.119007 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-config-data\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.119056 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a3869bdd-ef72-488e-8691-d4c6bc8d8118-ceph\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.119081 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.119106 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4340e44d-1160-461e-859a-efebef2f545b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.119145 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a3869bdd-ef72-488e-8691-d4c6bc8d8118-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.119181 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4cmr\" (UniqueName: \"kubernetes.io/projected/a3869bdd-ef72-488e-8691-d4c6bc8d8118-kube-api-access-g4cmr\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.119233 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3869bdd-ef72-488e-8691-d4c6bc8d8118-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.119757 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4340e44d-1160-461e-859a-efebef2f545b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.122335 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/a3869bdd-ef72-488e-8691-d4c6bc8d8118-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.124637 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a3869bdd-ef72-488e-8691-d4c6bc8d8118-ceph\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.124927 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.126834 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-scripts\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.127201 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.127526 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-config-data\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.127774 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-scripts\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.128510 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-config-data\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.130215 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3869bdd-ef72-488e-8691-d4c6bc8d8118-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.140127 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4cmr\" (UniqueName: \"kubernetes.io/projected/a3869bdd-ef72-488e-8691-d4c6bc8d8118-kube-api-access-g4cmr\") pod \"manila-share-share1-0\" (UID: \"a3869bdd-ef72-488e-8691-d4c6bc8d8118\") " pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.144303 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqczh\" (UniqueName: \"kubernetes.io/projected/4340e44d-1160-461e-859a-efebef2f545b-kube-api-access-gqczh\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.158527 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4340e44d-1160-461e-859a-efebef2f545b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"4340e44d-1160-461e-859a-efebef2f545b\") " pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.168379 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.220954 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60f96a82-8665-49b1-9f4e-8a3923e6d996-etc-machine-id\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221001 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-scripts\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221026 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-dns-svc\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221065 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-config\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221083 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf5sk\" (UniqueName: \"kubernetes.io/projected/60f96a82-8665-49b1-9f4e-8a3923e6d996-kube-api-access-qf5sk\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221099 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221134 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-sb\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221197 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjcbh\" (UniqueName: \"kubernetes.io/projected/5874184f-b246-4c70-a084-54cb58364d99-kube-api-access-sjcbh\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221221 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-config-data-custom\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221248 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-config-data\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221268 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60f96a82-8665-49b1-9f4e-8a3923e6d996-logs\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.221295 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-nb\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.236100 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.317150 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326234 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-config-data\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326277 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60f96a82-8665-49b1-9f4e-8a3923e6d996-logs\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326310 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-nb\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326364 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60f96a82-8665-49b1-9f4e-8a3923e6d996-etc-machine-id\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326383 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-scripts\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326403 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-dns-svc\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326437 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-config\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326454 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf5sk\" (UniqueName: \"kubernetes.io/projected/60f96a82-8665-49b1-9f4e-8a3923e6d996-kube-api-access-qf5sk\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326471 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326502 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-sb\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326561 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjcbh\" (UniqueName: \"kubernetes.io/projected/5874184f-b246-4c70-a084-54cb58364d99-kube-api-access-sjcbh\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.326587 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-config-data-custom\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.327856 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-config\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.328199 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60f96a82-8665-49b1-9f4e-8a3923e6d996-etc-machine-id\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.328413 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-dns-svc\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.328668 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-sb\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.329076 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60f96a82-8665-49b1-9f4e-8a3923e6d996-logs\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.330302 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-nb\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.332855 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.371869 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-scripts\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.371938 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-config-data\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.381939 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60f96a82-8665-49b1-9f4e-8a3923e6d996-config-data-custom\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.388674 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjcbh\" (UniqueName: \"kubernetes.io/projected/5874184f-b246-4c70-a084-54cb58364d99-kube-api-access-sjcbh\") pod \"dnsmasq-dns-77897597d9-bfqsd\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.399224 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf5sk\" (UniqueName: \"kubernetes.io/projected/60f96a82-8665-49b1-9f4e-8a3923e6d996-kube-api-access-qf5sk\") pod \"manila-api-0\" (UID: \"60f96a82-8665-49b1-9f4e-8a3923e6d996\") " pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.406476 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.660643 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.747126 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 05 09:01:46 crc kubenswrapper[4935]: I1005 09:01:46.997829 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 05 09:01:47 crc kubenswrapper[4935]: I1005 09:01:47.278078 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77897597d9-bfqsd"] Oct 05 09:01:47 crc kubenswrapper[4935]: I1005 09:01:47.290663 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 05 09:01:47 crc kubenswrapper[4935]: W1005 09:01:47.293146 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5874184f_b246_4c70_a084_54cb58364d99.slice/crio-a24360782244a22f8b10365b612ad476082f1a588aa206bfb5995b2cedd4ef21 WatchSource:0}: Error finding container a24360782244a22f8b10365b612ad476082f1a588aa206bfb5995b2cedd4ef21: Status 404 returned error can't find the container with id a24360782244a22f8b10365b612ad476082f1a588aa206bfb5995b2cedd4ef21 Oct 05 09:01:47 crc kubenswrapper[4935]: I1005 09:01:47.441432 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"4340e44d-1160-461e-859a-efebef2f545b","Type":"ContainerStarted","Data":"fcadf12e203a50a5489715b637128f376f9aeaf91f2358b0494d03f65025a5c0"} Oct 05 09:01:47 crc kubenswrapper[4935]: I1005 09:01:47.444677 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"60f96a82-8665-49b1-9f4e-8a3923e6d996","Type":"ContainerStarted","Data":"3ea42f039e4cc869ffb556c21a36055ef5fd0c57fe00e94d884668fc8e0a28ea"} Oct 05 09:01:47 crc kubenswrapper[4935]: I1005 09:01:47.456515 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" event={"ID":"5874184f-b246-4c70-a084-54cb58364d99","Type":"ContainerStarted","Data":"a24360782244a22f8b10365b612ad476082f1a588aa206bfb5995b2cedd4ef21"} Oct 05 09:01:47 crc kubenswrapper[4935]: I1005 09:01:47.462803 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a3869bdd-ef72-488e-8691-d4c6bc8d8118","Type":"ContainerStarted","Data":"e186cb9503fe1ce2f81321f7ed07bb817d8fc2ec35cb474a7a1bb22f043ce7a2"} Oct 05 09:01:48 crc kubenswrapper[4935]: I1005 09:01:48.475456 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"4340e44d-1160-461e-859a-efebef2f545b","Type":"ContainerStarted","Data":"362ee8210dde3c32d5ad16c6ee7c4e59a08613722dc91d9517e2e5eb69054983"} Oct 05 09:01:48 crc kubenswrapper[4935]: I1005 09:01:48.479734 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"60f96a82-8665-49b1-9f4e-8a3923e6d996","Type":"ContainerStarted","Data":"087b7ea477baac9c694cd6fd7d6bddea030a3069e4238578598c95b0d9b68044"} Oct 05 09:01:48 crc kubenswrapper[4935]: I1005 09:01:48.479785 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"60f96a82-8665-49b1-9f4e-8a3923e6d996","Type":"ContainerStarted","Data":"7d0f2bfeaecfcfc527865b28b24c04701584d53b51d7373d241703c874251acf"} Oct 05 09:01:48 crc kubenswrapper[4935]: I1005 09:01:48.479987 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 05 09:01:48 crc kubenswrapper[4935]: I1005 09:01:48.482732 4935 generic.go:334] "Generic (PLEG): container finished" podID="5874184f-b246-4c70-a084-54cb58364d99" containerID="7eded050f7c829fb706206eb3d787986bf82e99c7d502e52f6c933df168b433c" exitCode=0 Oct 05 09:01:48 crc kubenswrapper[4935]: I1005 09:01:48.482766 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" event={"ID":"5874184f-b246-4c70-a084-54cb58364d99","Type":"ContainerDied","Data":"7eded050f7c829fb706206eb3d787986bf82e99c7d502e52f6c933df168b433c"} Oct 05 09:01:48 crc kubenswrapper[4935]: I1005 09:01:48.568836 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=2.568809464 podStartE2EDuration="2.568809464s" podCreationTimestamp="2025-10-05 09:01:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:01:48.503958875 +0000 UTC m=+7742.386585355" watchObservedRunningTime="2025-10-05 09:01:48.568809464 +0000 UTC m=+7742.451435914" Oct 05 09:01:49 crc kubenswrapper[4935]: I1005 09:01:49.515585 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" event={"ID":"5874184f-b246-4c70-a084-54cb58364d99","Type":"ContainerStarted","Data":"6ef9384e6257505cac2fb061560b5e4c84bf8e01fefcf54d27925d73205b3c5d"} Oct 05 09:01:49 crc kubenswrapper[4935]: I1005 09:01:49.516080 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:49 crc kubenswrapper[4935]: I1005 09:01:49.523006 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"4340e44d-1160-461e-859a-efebef2f545b","Type":"ContainerStarted","Data":"5a9e5a39bb5fb5290c8cc5767741043d19b7ee2b645ed713aa972e28894f0c2d"} Oct 05 09:01:49 crc kubenswrapper[4935]: I1005 09:01:49.540156 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" podStartSLOduration=4.540138957 podStartE2EDuration="4.540138957s" podCreationTimestamp="2025-10-05 09:01:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:01:49.529600468 +0000 UTC m=+7743.412226948" watchObservedRunningTime="2025-10-05 09:01:49.540138957 +0000 UTC m=+7743.422765417" Oct 05 09:01:49 crc kubenswrapper[4935]: I1005 09:01:49.555153 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.996863664 podStartE2EDuration="4.555130405s" podCreationTimestamp="2025-10-05 09:01:45 +0000 UTC" firstStartedPulling="2025-10-05 09:01:46.80154354 +0000 UTC m=+7740.684170000" lastFinishedPulling="2025-10-05 09:01:47.359810281 +0000 UTC m=+7741.242436741" observedRunningTime="2025-10-05 09:01:49.551512449 +0000 UTC m=+7743.434138909" watchObservedRunningTime="2025-10-05 09:01:49.555130405 +0000 UTC m=+7743.437756875" Oct 05 09:01:50 crc kubenswrapper[4935]: I1005 09:01:50.777416 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:01:50 crc kubenswrapper[4935]: E1005 09:01:50.778005 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.050651 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-24rg4"] Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.064036 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-fl9t6"] Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.076489 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-fl9t6"] Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.086922 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-24rg4"] Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.100259 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-m8ddf"] Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.111063 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-m8ddf"] Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.595821 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a3869bdd-ef72-488e-8691-d4c6bc8d8118","Type":"ContainerStarted","Data":"47ddaa7e61aa4656fcd375d471db9c90f497b453d78539316c410a3d92c3b792"} Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.595876 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"a3869bdd-ef72-488e-8691-d4c6bc8d8118","Type":"ContainerStarted","Data":"34454c3d3c195ca8196b3a8502038b34e81932fa8f92bdefc112e906c8e06a96"} Oct 05 09:01:55 crc kubenswrapper[4935]: I1005 09:01:55.626097 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.513525628 podStartE2EDuration="10.626076561s" podCreationTimestamp="2025-10-05 09:01:45 +0000 UTC" firstStartedPulling="2025-10-05 09:01:47.301656249 +0000 UTC m=+7741.184282709" lastFinishedPulling="2025-10-05 09:01:54.414207182 +0000 UTC m=+7748.296833642" observedRunningTime="2025-10-05 09:01:55.622869246 +0000 UTC m=+7749.505495776" watchObservedRunningTime="2025-10-05 09:01:55.626076561 +0000 UTC m=+7749.508703031" Oct 05 09:01:56 crc kubenswrapper[4935]: I1005 09:01:56.169170 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 05 09:01:56 crc kubenswrapper[4935]: I1005 09:01:56.236720 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 05 09:01:56 crc kubenswrapper[4935]: I1005 09:01:56.664958 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:01:56 crc kubenswrapper[4935]: I1005 09:01:56.796984 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="183c385d-9a5a-484f-b4b4-b59f41a3af84" path="/var/lib/kubelet/pods/183c385d-9a5a-484f-b4b4-b59f41a3af84/volumes" Oct 05 09:01:56 crc kubenswrapper[4935]: I1005 09:01:56.797849 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="325ad13c-f69b-4c9d-9053-ad415c1f47a0" path="/var/lib/kubelet/pods/325ad13c-f69b-4c9d-9053-ad415c1f47a0/volumes" Oct 05 09:01:56 crc kubenswrapper[4935]: I1005 09:01:56.798636 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e" path="/var/lib/kubelet/pods/86b83239-95a1-4f6b-b6cf-72aa1bf3cb6e/volumes" Oct 05 09:01:56 crc kubenswrapper[4935]: I1005 09:01:56.799337 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b47d67949-vkqlg"] Oct 05 09:01:56 crc kubenswrapper[4935]: I1005 09:01:56.799604 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" podUID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" containerName="dnsmasq-dns" containerID="cri-o://1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa" gracePeriod=10 Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.360814 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.513081 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tgml\" (UniqueName: \"kubernetes.io/projected/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-kube-api-access-4tgml\") pod \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.513148 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-sb\") pod \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.513177 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-dns-svc\") pod \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.513302 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-config\") pod \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.513358 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-nb\") pod \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\" (UID: \"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f\") " Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.520737 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-kube-api-access-4tgml" (OuterVolumeSpecName: "kube-api-access-4tgml") pod "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" (UID: "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f"). InnerVolumeSpecName "kube-api-access-4tgml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.569537 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" (UID: "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.569756 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-config" (OuterVolumeSpecName: "config") pod "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" (UID: "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.572376 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" (UID: "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.582805 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" (UID: "133a2b2a-d0a9-475c-a990-1cfd88cd4c5f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.615196 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-config\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.615232 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.615247 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tgml\" (UniqueName: \"kubernetes.io/projected/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-kube-api-access-4tgml\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.615258 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.615267 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.618974 4935 generic.go:334] "Generic (PLEG): container finished" podID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" containerID="1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa" exitCode=0 Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.619178 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" event={"ID":"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f","Type":"ContainerDied","Data":"1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa"} Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.619310 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" event={"ID":"133a2b2a-d0a9-475c-a990-1cfd88cd4c5f","Type":"ContainerDied","Data":"4a84f78519c144f76ee595e0be1bc05d3f6bb5802c1bb52960dce02971cd03e4"} Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.619426 4935 scope.go:117] "RemoveContainer" containerID="1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.619706 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b47d67949-vkqlg" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.664949 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b47d67949-vkqlg"] Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.672755 4935 scope.go:117] "RemoveContainer" containerID="77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.672851 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b47d67949-vkqlg"] Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.695805 4935 scope.go:117] "RemoveContainer" containerID="1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa" Oct 05 09:01:57 crc kubenswrapper[4935]: E1005 09:01:57.696400 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa\": container with ID starting with 1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa not found: ID does not exist" containerID="1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.696440 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa"} err="failed to get container status \"1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa\": rpc error: code = NotFound desc = could not find container \"1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa\": container with ID starting with 1e41ed241feea86a10ed45b57512080f2fd56c186244901e8356a1bc45c2a5fa not found: ID does not exist" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.696467 4935 scope.go:117] "RemoveContainer" containerID="77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320" Oct 05 09:01:57 crc kubenswrapper[4935]: E1005 09:01:57.697486 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320\": container with ID starting with 77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320 not found: ID does not exist" containerID="77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320" Oct 05 09:01:57 crc kubenswrapper[4935]: I1005 09:01:57.697640 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320"} err="failed to get container status \"77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320\": rpc error: code = NotFound desc = could not find container \"77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320\": container with ID starting with 77ca2702942c5e61d8ce756bb514dc0907da199bd3cbaa6e490858d237f1e320 not found: ID does not exist" Oct 05 09:01:58 crc kubenswrapper[4935]: I1005 09:01:58.809548 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" path="/var/lib/kubelet/pods/133a2b2a-d0a9-475c-a990-1cfd88cd4c5f/volumes" Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.474491 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.475409 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="ceilometer-central-agent" containerID="cri-o://b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111" gracePeriod=30 Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.475565 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="sg-core" containerID="cri-o://51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb" gracePeriod=30 Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.475579 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="ceilometer-notification-agent" containerID="cri-o://73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761" gracePeriod=30 Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.475494 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="proxy-httpd" containerID="cri-o://09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859" gracePeriod=30 Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.661877 4935 generic.go:334] "Generic (PLEG): container finished" podID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerID="09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859" exitCode=0 Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.662051 4935 generic.go:334] "Generic (PLEG): container finished" podID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerID="51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb" exitCode=2 Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.661927 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerDied","Data":"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859"} Oct 05 09:01:59 crc kubenswrapper[4935]: I1005 09:01:59.662098 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerDied","Data":"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb"} Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.421740 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.594556 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-sg-core-conf-yaml\") pod \"0b167df5-0f56-489f-bb85-d48128d3ba07\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.594612 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-scripts\") pod \"0b167df5-0f56-489f-bb85-d48128d3ba07\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.594667 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-log-httpd\") pod \"0b167df5-0f56-489f-bb85-d48128d3ba07\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.594702 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d822\" (UniqueName: \"kubernetes.io/projected/0b167df5-0f56-489f-bb85-d48128d3ba07-kube-api-access-7d822\") pod \"0b167df5-0f56-489f-bb85-d48128d3ba07\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.594755 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-config-data\") pod \"0b167df5-0f56-489f-bb85-d48128d3ba07\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.594779 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-combined-ca-bundle\") pod \"0b167df5-0f56-489f-bb85-d48128d3ba07\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.594804 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-run-httpd\") pod \"0b167df5-0f56-489f-bb85-d48128d3ba07\" (UID: \"0b167df5-0f56-489f-bb85-d48128d3ba07\") " Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.595705 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0b167df5-0f56-489f-bb85-d48128d3ba07" (UID: "0b167df5-0f56-489f-bb85-d48128d3ba07"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.601727 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0b167df5-0f56-489f-bb85-d48128d3ba07" (UID: "0b167df5-0f56-489f-bb85-d48128d3ba07"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.603561 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-scripts" (OuterVolumeSpecName: "scripts") pod "0b167df5-0f56-489f-bb85-d48128d3ba07" (UID: "0b167df5-0f56-489f-bb85-d48128d3ba07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.610540 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b167df5-0f56-489f-bb85-d48128d3ba07-kube-api-access-7d822" (OuterVolumeSpecName: "kube-api-access-7d822") pod "0b167df5-0f56-489f-bb85-d48128d3ba07" (UID: "0b167df5-0f56-489f-bb85-d48128d3ba07"). InnerVolumeSpecName "kube-api-access-7d822". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.636005 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0b167df5-0f56-489f-bb85-d48128d3ba07" (UID: "0b167df5-0f56-489f-bb85-d48128d3ba07"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.687749 4935 generic.go:334] "Generic (PLEG): container finished" podID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerID="73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761" exitCode=0 Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.687798 4935 generic.go:334] "Generic (PLEG): container finished" podID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerID="b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111" exitCode=0 Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.687827 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerDied","Data":"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761"} Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.687868 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerDied","Data":"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111"} Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.687878 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b167df5-0f56-489f-bb85-d48128d3ba07","Type":"ContainerDied","Data":"b9a9e6ffad92a603742435ad66a5cb131ea39e4abf9a5cad03d030b0f531d798"} Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.687914 4935 scope.go:117] "RemoveContainer" containerID="09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.688043 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.696853 4935 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.697291 4935 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.697306 4935 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.697317 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d822\" (UniqueName: \"kubernetes.io/projected/0b167df5-0f56-489f-bb85-d48128d3ba07-kube-api-access-7d822\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.697333 4935 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b167df5-0f56-489f-bb85-d48128d3ba07-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.708982 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b167df5-0f56-489f-bb85-d48128d3ba07" (UID: "0b167df5-0f56-489f-bb85-d48128d3ba07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.716884 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-config-data" (OuterVolumeSpecName: "config-data") pod "0b167df5-0f56-489f-bb85-d48128d3ba07" (UID: "0b167df5-0f56-489f-bb85-d48128d3ba07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.798087 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.798121 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b167df5-0f56-489f-bb85-d48128d3ba07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.822254 4935 scope.go:117] "RemoveContainer" containerID="51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.850721 4935 scope.go:117] "RemoveContainer" containerID="73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.873669 4935 scope.go:117] "RemoveContainer" containerID="b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.903774 4935 scope.go:117] "RemoveContainer" containerID="09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859" Oct 05 09:02:00 crc kubenswrapper[4935]: E1005 09:02:00.904692 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859\": container with ID starting with 09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859 not found: ID does not exist" containerID="09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.904724 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859"} err="failed to get container status \"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859\": rpc error: code = NotFound desc = could not find container \"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859\": container with ID starting with 09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859 not found: ID does not exist" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.904747 4935 scope.go:117] "RemoveContainer" containerID="51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb" Oct 05 09:02:00 crc kubenswrapper[4935]: E1005 09:02:00.905226 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb\": container with ID starting with 51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb not found: ID does not exist" containerID="51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.905273 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb"} err="failed to get container status \"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb\": rpc error: code = NotFound desc = could not find container \"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb\": container with ID starting with 51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb not found: ID does not exist" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.905304 4935 scope.go:117] "RemoveContainer" containerID="73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761" Oct 05 09:02:00 crc kubenswrapper[4935]: E1005 09:02:00.905785 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761\": container with ID starting with 73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761 not found: ID does not exist" containerID="73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.905864 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761"} err="failed to get container status \"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761\": rpc error: code = NotFound desc = could not find container \"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761\": container with ID starting with 73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761 not found: ID does not exist" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.905937 4935 scope.go:117] "RemoveContainer" containerID="b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111" Oct 05 09:02:00 crc kubenswrapper[4935]: E1005 09:02:00.906414 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111\": container with ID starting with b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111 not found: ID does not exist" containerID="b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.906453 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111"} err="failed to get container status \"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111\": rpc error: code = NotFound desc = could not find container \"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111\": container with ID starting with b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111 not found: ID does not exist" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.906470 4935 scope.go:117] "RemoveContainer" containerID="09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.906938 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859"} err="failed to get container status \"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859\": rpc error: code = NotFound desc = could not find container \"09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859\": container with ID starting with 09eb85d216554d72585cbf33937f5e44440c86c5781aa40dc373c40ece7f8859 not found: ID does not exist" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.906963 4935 scope.go:117] "RemoveContainer" containerID="51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.907528 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb"} err="failed to get container status \"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb\": rpc error: code = NotFound desc = could not find container \"51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb\": container with ID starting with 51e4860a8d4b7fd90ede1e3ee9afeae05cd5d490b49ef650f907f3114d520cbb not found: ID does not exist" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.907574 4935 scope.go:117] "RemoveContainer" containerID="73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.908010 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761"} err="failed to get container status \"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761\": rpc error: code = NotFound desc = could not find container \"73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761\": container with ID starting with 73dccc0768f8cc2485405108a0c988016b84da09e01dd989f382594af4b7b761 not found: ID does not exist" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.908051 4935 scope.go:117] "RemoveContainer" containerID="b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111" Oct 05 09:02:00 crc kubenswrapper[4935]: I1005 09:02:00.908670 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111"} err="failed to get container status \"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111\": rpc error: code = NotFound desc = could not find container \"b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111\": container with ID starting with b54d3ad2b04faf5d7c50f016d910537eba16bc1744c1cf8e689c445809952111 not found: ID does not exist" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.016823 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.026052 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.058746 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:02:01 crc kubenswrapper[4935]: E1005 09:02:01.059182 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="ceilometer-central-agent" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059199 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="ceilometer-central-agent" Oct 05 09:02:01 crc kubenswrapper[4935]: E1005 09:02:01.059218 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="ceilometer-notification-agent" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059225 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="ceilometer-notification-agent" Oct 05 09:02:01 crc kubenswrapper[4935]: E1005 09:02:01.059241 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="proxy-httpd" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059247 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="proxy-httpd" Oct 05 09:02:01 crc kubenswrapper[4935]: E1005 09:02:01.059273 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" containerName="init" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059279 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" containerName="init" Oct 05 09:02:01 crc kubenswrapper[4935]: E1005 09:02:01.059291 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" containerName="dnsmasq-dns" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059307 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" containerName="dnsmasq-dns" Oct 05 09:02:01 crc kubenswrapper[4935]: E1005 09:02:01.059321 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="sg-core" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059326 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="sg-core" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059505 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="proxy-httpd" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059527 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="sg-core" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059537 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="ceilometer-notification-agent" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059545 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" containerName="ceilometer-central-agent" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.059556 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="133a2b2a-d0a9-475c-a990-1cfd88cd4c5f" containerName="dnsmasq-dns" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.061417 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.067559 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.067786 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.076124 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.205993 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-config-data\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.206050 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.206185 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndjzh\" (UniqueName: \"kubernetes.io/projected/be33a62f-6a3c-4893-8a04-c67b5c08159f-kube-api-access-ndjzh\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.206283 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be33a62f-6a3c-4893-8a04-c67b5c08159f-log-httpd\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.206612 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be33a62f-6a3c-4893-8a04-c67b5c08159f-run-httpd\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.206676 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.207047 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-scripts\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.308663 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-scripts\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.308797 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-config-data\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.308819 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.309175 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndjzh\" (UniqueName: \"kubernetes.io/projected/be33a62f-6a3c-4893-8a04-c67b5c08159f-kube-api-access-ndjzh\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.310881 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be33a62f-6a3c-4893-8a04-c67b5c08159f-log-httpd\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.311123 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be33a62f-6a3c-4893-8a04-c67b5c08159f-run-httpd\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.311176 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.311685 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be33a62f-6a3c-4893-8a04-c67b5c08159f-log-httpd\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.312071 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be33a62f-6a3c-4893-8a04-c67b5c08159f-run-httpd\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.313130 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-scripts\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.314186 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.314678 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.322226 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be33a62f-6a3c-4893-8a04-c67b5c08159f-config-data\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.324464 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndjzh\" (UniqueName: \"kubernetes.io/projected/be33a62f-6a3c-4893-8a04-c67b5c08159f-kube-api-access-ndjzh\") pod \"ceilometer-0\" (UID: \"be33a62f-6a3c-4893-8a04-c67b5c08159f\") " pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.359152 4935 scope.go:117] "RemoveContainer" containerID="0cbc289339fa8b5c1686b974fdd9fba076ddf5bc6fbb7f9de012a0846900857f" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.377694 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.385732 4935 scope.go:117] "RemoveContainer" containerID="3a6039b9da6eb593a8bf1904f927686f79a14ab313267d01c5f9232650f89fae" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.412457 4935 scope.go:117] "RemoveContainer" containerID="242b46b82049ae686f91fa3c5c4343f37219bc3f3ea5872b48c328caddfe92bc" Oct 05 09:02:01 crc kubenswrapper[4935]: I1005 09:02:01.910686 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 09:02:02 crc kubenswrapper[4935]: I1005 09:02:02.738188 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be33a62f-6a3c-4893-8a04-c67b5c08159f","Type":"ContainerStarted","Data":"84bc399972bafe11703ffd2630175d8fb4626041e18b6f935e944a3fa45c9271"} Oct 05 09:02:02 crc kubenswrapper[4935]: I1005 09:02:02.738439 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be33a62f-6a3c-4893-8a04-c67b5c08159f","Type":"ContainerStarted","Data":"a7dfa0f4639bfbe5d783d0306dcf93b385152a6c615f414c41c23bff798ad8c6"} Oct 05 09:02:02 crc kubenswrapper[4935]: I1005 09:02:02.798330 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b167df5-0f56-489f-bb85-d48128d3ba07" path="/var/lib/kubelet/pods/0b167df5-0f56-489f-bb85-d48128d3ba07/volumes" Oct 05 09:02:03 crc kubenswrapper[4935]: I1005 09:02:03.777321 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:02:03 crc kubenswrapper[4935]: E1005 09:02:03.778349 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:02:04 crc kubenswrapper[4935]: I1005 09:02:04.758331 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be33a62f-6a3c-4893-8a04-c67b5c08159f","Type":"ContainerStarted","Data":"399627220dff9ec9abe47debe9ac3c3cc54ce56d197b4db7bfe74002dccb2f9e"} Oct 05 09:02:05 crc kubenswrapper[4935]: I1005 09:02:05.048617 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0f81-account-create-brq6v"] Oct 05 09:02:05 crc kubenswrapper[4935]: I1005 09:02:05.059921 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0f81-account-create-brq6v"] Oct 05 09:02:05 crc kubenswrapper[4935]: I1005 09:02:05.768085 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be33a62f-6a3c-4893-8a04-c67b5c08159f","Type":"ContainerStarted","Data":"d6f7d21a972bfcb2b67dc350213c9ca8886150f5b65fd3d62238973d64d2e7f8"} Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.037174 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-83b2-account-create-5mvxp"] Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.052356 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f522-account-create-rqblm"] Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.061744 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-83b2-account-create-5mvxp"] Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.072800 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f522-account-create-rqblm"] Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.789155 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f71a336-51f3-4ad9-8dab-bc801ac5c0d9" path="/var/lib/kubelet/pods/0f71a336-51f3-4ad9-8dab-bc801ac5c0d9/volumes" Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.794358 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c993f0b-5b16-4710-89ce-e5e1881479bb" path="/var/lib/kubelet/pods/4c993f0b-5b16-4710-89ce-e5e1881479bb/volumes" Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.798056 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca9810c5-05cd-440e-8b35-54940b7beb4f" path="/var/lib/kubelet/pods/ca9810c5-05cd-440e-8b35-54940b7beb4f/volumes" Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.798805 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be33a62f-6a3c-4893-8a04-c67b5c08159f","Type":"ContainerStarted","Data":"79e9384efa47952c597670106686fc356e50d42d5d18963a2e18b7e0a1152c36"} Oct 05 09:02:06 crc kubenswrapper[4935]: I1005 09:02:06.851212 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.251018806 podStartE2EDuration="5.851190338s" podCreationTimestamp="2025-10-05 09:02:01 +0000 UTC" firstStartedPulling="2025-10-05 09:02:01.908863895 +0000 UTC m=+7755.791490395" lastFinishedPulling="2025-10-05 09:02:06.509035467 +0000 UTC m=+7760.391661927" observedRunningTime="2025-10-05 09:02:06.84561309 +0000 UTC m=+7760.728239550" watchObservedRunningTime="2025-10-05 09:02:06.851190338 +0000 UTC m=+7760.733816798" Oct 05 09:02:07 crc kubenswrapper[4935]: I1005 09:02:07.744523 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 05 09:02:07 crc kubenswrapper[4935]: I1005 09:02:07.785303 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 05 09:02:07 crc kubenswrapper[4935]: I1005 09:02:07.791800 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 09:02:08 crc kubenswrapper[4935]: I1005 09:02:08.085764 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 05 09:02:14 crc kubenswrapper[4935]: I1005 09:02:14.781038 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:02:14 crc kubenswrapper[4935]: E1005 09:02:14.806545 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:02:23 crc kubenswrapper[4935]: I1005 09:02:23.041536 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kbvgl"] Oct 05 09:02:23 crc kubenswrapper[4935]: I1005 09:02:23.062445 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kbvgl"] Oct 05 09:02:24 crc kubenswrapper[4935]: I1005 09:02:24.789575 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ce69a4-c3f0-4585-893a-8da9ff928ce1" path="/var/lib/kubelet/pods/14ce69a4-c3f0-4585-893a-8da9ff928ce1/volumes" Oct 05 09:02:25 crc kubenswrapper[4935]: I1005 09:02:25.778230 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:02:25 crc kubenswrapper[4935]: E1005 09:02:25.778630 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:02:31 crc kubenswrapper[4935]: I1005 09:02:31.382964 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 09:02:38 crc kubenswrapper[4935]: I1005 09:02:38.049051 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9j8m"] Oct 05 09:02:38 crc kubenswrapper[4935]: I1005 09:02:38.058486 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9j8m"] Oct 05 09:02:38 crc kubenswrapper[4935]: I1005 09:02:38.789348 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="710e234a-a567-428d-855a-dc5c5a695860" path="/var/lib/kubelet/pods/710e234a-a567-428d-855a-dc5c5a695860/volumes" Oct 05 09:02:39 crc kubenswrapper[4935]: I1005 09:02:39.034868 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z52d2"] Oct 05 09:02:39 crc kubenswrapper[4935]: I1005 09:02:39.042527 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z52d2"] Oct 05 09:02:40 crc kubenswrapper[4935]: I1005 09:02:40.799537 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93d805da-2c17-4b2a-ad2f-609a03683de9" path="/var/lib/kubelet/pods/93d805da-2c17-4b2a-ad2f-609a03683de9/volumes" Oct 05 09:02:41 crc kubenswrapper[4935]: I1005 09:02:41.777805 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:02:41 crc kubenswrapper[4935]: E1005 09:02:41.778244 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.777820 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:02:54 crc kubenswrapper[4935]: E1005 09:02:54.778711 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.799928 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7847dbb94c-bjcc6"] Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.802075 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.808493 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.815855 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7847dbb94c-bjcc6"] Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.855207 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-sb\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.855345 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nkjn\" (UniqueName: \"kubernetes.io/projected/35e433a5-2009-4c4e-9e85-49810811b71c-kube-api-access-5nkjn\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.855385 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-openstack-cell1\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.855440 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-dns-svc\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.855540 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-nb\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.855602 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-config\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.957854 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-nb\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.957945 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-config\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.958042 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-sb\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.958102 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nkjn\" (UniqueName: \"kubernetes.io/projected/35e433a5-2009-4c4e-9e85-49810811b71c-kube-api-access-5nkjn\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.958136 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-openstack-cell1\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.958175 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-dns-svc\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.958806 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-nb\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.959115 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-config\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.959161 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-sb\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.959164 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-dns-svc\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.959745 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-openstack-cell1\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:54 crc kubenswrapper[4935]: I1005 09:02:54.986715 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nkjn\" (UniqueName: \"kubernetes.io/projected/35e433a5-2009-4c4e-9e85-49810811b71c-kube-api-access-5nkjn\") pod \"dnsmasq-dns-7847dbb94c-bjcc6\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:55 crc kubenswrapper[4935]: I1005 09:02:55.125217 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:55 crc kubenswrapper[4935]: I1005 09:02:55.593569 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7847dbb94c-bjcc6"] Oct 05 09:02:56 crc kubenswrapper[4935]: I1005 09:02:56.314649 4935 generic.go:334] "Generic (PLEG): container finished" podID="35e433a5-2009-4c4e-9e85-49810811b71c" containerID="f1e7be0feb6a4890002374f8519d843d4b2988ffe5740599f57bc7aadc12d707" exitCode=0 Oct 05 09:02:56 crc kubenswrapper[4935]: I1005 09:02:56.314818 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" event={"ID":"35e433a5-2009-4c4e-9e85-49810811b71c","Type":"ContainerDied","Data":"f1e7be0feb6a4890002374f8519d843d4b2988ffe5740599f57bc7aadc12d707"} Oct 05 09:02:56 crc kubenswrapper[4935]: I1005 09:02:56.316341 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" event={"ID":"35e433a5-2009-4c4e-9e85-49810811b71c","Type":"ContainerStarted","Data":"4f4791727142037fad5474831d0e480bacf91095dd773318792166a9ec857934"} Oct 05 09:02:57 crc kubenswrapper[4935]: I1005 09:02:57.325432 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" event={"ID":"35e433a5-2009-4c4e-9e85-49810811b71c","Type":"ContainerStarted","Data":"0d3d3e6ff01b78a931663c6251387287897e9e19e2021670230cbfb50b83e32c"} Oct 05 09:02:57 crc kubenswrapper[4935]: I1005 09:02:57.326037 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:02:58 crc kubenswrapper[4935]: I1005 09:02:58.029165 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" podStartSLOduration=4.029145952 podStartE2EDuration="4.029145952s" podCreationTimestamp="2025-10-05 09:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:02:57.355537453 +0000 UTC m=+7811.238163923" watchObservedRunningTime="2025-10-05 09:02:58.029145952 +0000 UTC m=+7811.911772412" Oct 05 09:02:58 crc kubenswrapper[4935]: I1005 09:02:58.034656 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mvdz8"] Oct 05 09:02:58 crc kubenswrapper[4935]: I1005 09:02:58.042833 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mvdz8"] Oct 05 09:02:58 crc kubenswrapper[4935]: I1005 09:02:58.793255 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1077b73f-70fa-4968-9ddb-4e2148bdea80" path="/var/lib/kubelet/pods/1077b73f-70fa-4968-9ddb-4e2148bdea80/volumes" Oct 05 09:03:01 crc kubenswrapper[4935]: I1005 09:03:01.567386 4935 scope.go:117] "RemoveContainer" containerID="38febbbc3e3f399bccafc6b2ffb7ae636a7298b1a6d204b13dac4aa8b98357a3" Oct 05 09:03:01 crc kubenswrapper[4935]: I1005 09:03:01.598696 4935 scope.go:117] "RemoveContainer" containerID="c6f2d5876ec0bd8c72fddf763c71a407f2d376afccab2664eed0b703fcc0e468" Oct 05 09:03:01 crc kubenswrapper[4935]: I1005 09:03:01.696011 4935 scope.go:117] "RemoveContainer" containerID="e36244e8ba12537dfd22c4da79722910b3092fe0975fdf32c75ed4b144f4b0fa" Oct 05 09:03:01 crc kubenswrapper[4935]: I1005 09:03:01.779286 4935 scope.go:117] "RemoveContainer" containerID="b6fe144670d2f5db06ced874e32576eeaa621ab9f5420f3049766466efe7da56" Oct 05 09:03:01 crc kubenswrapper[4935]: I1005 09:03:01.830015 4935 scope.go:117] "RemoveContainer" containerID="13e3a91d48287ce623359e4ccfaba56866dc589067bb8fcf5ddb2de262b1282b" Oct 05 09:03:01 crc kubenswrapper[4935]: I1005 09:03:01.869568 4935 scope.go:117] "RemoveContainer" containerID="01372cc20d63a99b396530eb9913597e8f692740c875d3e6ab7f8e2230e7568c" Oct 05 09:03:01 crc kubenswrapper[4935]: I1005 09:03:01.892102 4935 scope.go:117] "RemoveContainer" containerID="e4d6661be0737c7116633e7c4d2f3eecc46901ad9577ad1a4a7f305acc479636" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.127167 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.210128 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77897597d9-bfqsd"] Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.210734 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" podUID="5874184f-b246-4c70-a084-54cb58364d99" containerName="dnsmasq-dns" containerID="cri-o://6ef9384e6257505cac2fb061560b5e4c84bf8e01fefcf54d27925d73205b3c5d" gracePeriod=10 Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.425217 4935 generic.go:334] "Generic (PLEG): container finished" podID="5874184f-b246-4c70-a084-54cb58364d99" containerID="6ef9384e6257505cac2fb061560b5e4c84bf8e01fefcf54d27925d73205b3c5d" exitCode=0 Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.425257 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" event={"ID":"5874184f-b246-4c70-a084-54cb58364d99","Type":"ContainerDied","Data":"6ef9384e6257505cac2fb061560b5e4c84bf8e01fefcf54d27925d73205b3c5d"} Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.479048 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8644c4685-dgr6h"] Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.490375 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.493268 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8644c4685-dgr6h"] Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.494472 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-networker" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.598002 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8644c4685-dgr6h"] Oct 05 09:03:05 crc kubenswrapper[4935]: E1005 09:03:05.598732 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-6w8bz openstack-cell1 openstack-networker ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-8644c4685-dgr6h" podUID="adc0254d-55a6-470b-a7f4-0f5f972e6963" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.626722 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7dbb57f589-r2bjx"] Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.628514 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.632511 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-config\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.632581 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-networker\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.632675 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-dns-svc\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.632719 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-cell1\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.632802 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-sb\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.632881 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-nb\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.632916 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w8bz\" (UniqueName: \"kubernetes.io/projected/adc0254d-55a6-470b-a7f4-0f5f972e6963-kube-api-access-6w8bz\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.650301 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dbb57f589-r2bjx"] Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734185 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkphf\" (UniqueName: \"kubernetes.io/projected/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-kube-api-access-xkphf\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734531 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734569 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-sb\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734600 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-nb\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734635 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w8bz\" (UniqueName: \"kubernetes.io/projected/adc0254d-55a6-470b-a7f4-0f5f972e6963-kube-api-access-6w8bz\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734673 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-config\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734724 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-networker\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734756 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-config\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734790 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-openstack-networker\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734829 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-dns-svc\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734865 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-dns-svc\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734909 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-cell1\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734959 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-openstack-cell1\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.734998 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.735933 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-sb\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.736529 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-nb\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.737395 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-config\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.739176 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-networker\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.739881 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-dns-svc\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.740604 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-cell1\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.762663 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w8bz\" (UniqueName: \"kubernetes.io/projected/adc0254d-55a6-470b-a7f4-0f5f972e6963-kube-api-access-6w8bz\") pod \"dnsmasq-dns-8644c4685-dgr6h\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.777053 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:03:05 crc kubenswrapper[4935]: E1005 09:03:05.777419 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.828226 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.836562 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-config\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.836804 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-openstack-networker\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.837021 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-dns-svc\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.837213 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-openstack-cell1\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.837339 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.837439 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkphf\" (UniqueName: \"kubernetes.io/projected/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-kube-api-access-xkphf\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.837517 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.837598 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-config\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.838361 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.838698 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-openstack-cell1\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.838968 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-dns-svc\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.839515 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-openstack-networker\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.839523 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.858769 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkphf\" (UniqueName: \"kubernetes.io/projected/37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e-kube-api-access-xkphf\") pod \"dnsmasq-dns-7dbb57f589-r2bjx\" (UID: \"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e\") " pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.938813 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-nb\") pod \"5874184f-b246-4c70-a084-54cb58364d99\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.938882 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-sb\") pod \"5874184f-b246-4c70-a084-54cb58364d99\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.938917 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-config\") pod \"5874184f-b246-4c70-a084-54cb58364d99\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.938994 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-dns-svc\") pod \"5874184f-b246-4c70-a084-54cb58364d99\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.939059 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjcbh\" (UniqueName: \"kubernetes.io/projected/5874184f-b246-4c70-a084-54cb58364d99-kube-api-access-sjcbh\") pod \"5874184f-b246-4c70-a084-54cb58364d99\" (UID: \"5874184f-b246-4c70-a084-54cb58364d99\") " Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.942260 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5874184f-b246-4c70-a084-54cb58364d99-kube-api-access-sjcbh" (OuterVolumeSpecName: "kube-api-access-sjcbh") pod "5874184f-b246-4c70-a084-54cb58364d99" (UID: "5874184f-b246-4c70-a084-54cb58364d99"). InnerVolumeSpecName "kube-api-access-sjcbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.952919 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:05 crc kubenswrapper[4935]: I1005 09:03:05.994450 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5874184f-b246-4c70-a084-54cb58364d99" (UID: "5874184f-b246-4c70-a084-54cb58364d99"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.005456 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-config" (OuterVolumeSpecName: "config") pod "5874184f-b246-4c70-a084-54cb58364d99" (UID: "5874184f-b246-4c70-a084-54cb58364d99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.005987 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5874184f-b246-4c70-a084-54cb58364d99" (UID: "5874184f-b246-4c70-a084-54cb58364d99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.008083 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5874184f-b246-4c70-a084-54cb58364d99" (UID: "5874184f-b246-4c70-a084-54cb58364d99"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.041680 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.041715 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.041728 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-config\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.041741 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5874184f-b246-4c70-a084-54cb58364d99-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.041755 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjcbh\" (UniqueName: \"kubernetes.io/projected/5874184f-b246-4c70-a084-54cb58364d99-kube-api-access-sjcbh\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.428440 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dbb57f589-r2bjx"] Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.445978 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.446051 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.445971 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77897597d9-bfqsd" event={"ID":"5874184f-b246-4c70-a084-54cb58364d99","Type":"ContainerDied","Data":"a24360782244a22f8b10365b612ad476082f1a588aa206bfb5995b2cedd4ef21"} Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.446198 4935 scope.go:117] "RemoveContainer" containerID="6ef9384e6257505cac2fb061560b5e4c84bf8e01fefcf54d27925d73205b3c5d" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.644233 4935 scope.go:117] "RemoveContainer" containerID="7eded050f7c829fb706206eb3d787986bf82e99c7d502e52f6c933df168b433c" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.710850 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.775828 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77897597d9-bfqsd"] Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.826739 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77897597d9-bfqsd"] Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.869502 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w8bz\" (UniqueName: \"kubernetes.io/projected/adc0254d-55a6-470b-a7f4-0f5f972e6963-kube-api-access-6w8bz\") pod \"adc0254d-55a6-470b-a7f4-0f5f972e6963\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.869595 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-nb\") pod \"adc0254d-55a6-470b-a7f4-0f5f972e6963\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.869650 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-sb\") pod \"adc0254d-55a6-470b-a7f4-0f5f972e6963\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.869702 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-networker\") pod \"adc0254d-55a6-470b-a7f4-0f5f972e6963\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.869775 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-config\") pod \"adc0254d-55a6-470b-a7f4-0f5f972e6963\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.869790 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-dns-svc\") pod \"adc0254d-55a6-470b-a7f4-0f5f972e6963\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.869807 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-cell1\") pod \"adc0254d-55a6-470b-a7f4-0f5f972e6963\" (UID: \"adc0254d-55a6-470b-a7f4-0f5f972e6963\") " Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.871011 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "adc0254d-55a6-470b-a7f4-0f5f972e6963" (UID: "adc0254d-55a6-470b-a7f4-0f5f972e6963"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.871484 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-config" (OuterVolumeSpecName: "config") pod "adc0254d-55a6-470b-a7f4-0f5f972e6963" (UID: "adc0254d-55a6-470b-a7f4-0f5f972e6963"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.871606 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "adc0254d-55a6-470b-a7f4-0f5f972e6963" (UID: "adc0254d-55a6-470b-a7f4-0f5f972e6963"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.871794 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "adc0254d-55a6-470b-a7f4-0f5f972e6963" (UID: "adc0254d-55a6-470b-a7f4-0f5f972e6963"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.871960 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "adc0254d-55a6-470b-a7f4-0f5f972e6963" (UID: "adc0254d-55a6-470b-a7f4-0f5f972e6963"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.872283 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "adc0254d-55a6-470b-a7f4-0f5f972e6963" (UID: "adc0254d-55a6-470b-a7f4-0f5f972e6963"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.883367 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc0254d-55a6-470b-a7f4-0f5f972e6963-kube-api-access-6w8bz" (OuterVolumeSpecName: "kube-api-access-6w8bz") pod "adc0254d-55a6-470b-a7f4-0f5f972e6963" (UID: "adc0254d-55a6-470b-a7f4-0f5f972e6963"). InnerVolumeSpecName "kube-api-access-6w8bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.973604 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-networker\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.973638 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-config\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.973652 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.973665 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.973674 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w8bz\" (UniqueName: \"kubernetes.io/projected/adc0254d-55a6-470b-a7f4-0f5f972e6963-kube-api-access-6w8bz\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.973683 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:06 crc kubenswrapper[4935]: I1005 09:03:06.973692 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adc0254d-55a6-470b-a7f4-0f5f972e6963-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:07 crc kubenswrapper[4935]: I1005 09:03:07.460765 4935 generic.go:334] "Generic (PLEG): container finished" podID="37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e" containerID="d3372fb2617404a3fe334e6ff2aa6b0e93567eb8dd0efd2d7da7750e725a433a" exitCode=0 Oct 05 09:03:07 crc kubenswrapper[4935]: I1005 09:03:07.461087 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8644c4685-dgr6h" Oct 05 09:03:07 crc kubenswrapper[4935]: I1005 09:03:07.465243 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" event={"ID":"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e","Type":"ContainerDied","Data":"d3372fb2617404a3fe334e6ff2aa6b0e93567eb8dd0efd2d7da7750e725a433a"} Oct 05 09:03:07 crc kubenswrapper[4935]: I1005 09:03:07.465308 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" event={"ID":"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e","Type":"ContainerStarted","Data":"c7f618e859cf26d0548902383b254225cc35e52b3cf06bf95f507bbeda48fa5c"} Oct 05 09:03:07 crc kubenswrapper[4935]: I1005 09:03:07.613964 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8644c4685-dgr6h"] Oct 05 09:03:07 crc kubenswrapper[4935]: I1005 09:03:07.629014 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8644c4685-dgr6h"] Oct 05 09:03:08 crc kubenswrapper[4935]: I1005 09:03:08.472268 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" event={"ID":"37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e","Type":"ContainerStarted","Data":"3864ff3ce6f93d25ca180545721c92750dbeb62976f26ce4e075785420bfaf60"} Oct 05 09:03:08 crc kubenswrapper[4935]: I1005 09:03:08.472446 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:08 crc kubenswrapper[4935]: I1005 09:03:08.502668 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" podStartSLOduration=3.502645814 podStartE2EDuration="3.502645814s" podCreationTimestamp="2025-10-05 09:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:03:08.492458904 +0000 UTC m=+7822.375085424" watchObservedRunningTime="2025-10-05 09:03:08.502645814 +0000 UTC m=+7822.385272294" Oct 05 09:03:08 crc kubenswrapper[4935]: I1005 09:03:08.790411 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5874184f-b246-4c70-a084-54cb58364d99" path="/var/lib/kubelet/pods/5874184f-b246-4c70-a084-54cb58364d99/volumes" Oct 05 09:03:08 crc kubenswrapper[4935]: I1005 09:03:08.791290 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adc0254d-55a6-470b-a7f4-0f5f972e6963" path="/var/lib/kubelet/pods/adc0254d-55a6-470b-a7f4-0f5f972e6963/volumes" Oct 05 09:03:15 crc kubenswrapper[4935]: I1005 09:03:15.955166 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7dbb57f589-r2bjx" Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.050934 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7847dbb94c-bjcc6"] Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.051222 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" podUID="35e433a5-2009-4c4e-9e85-49810811b71c" containerName="dnsmasq-dns" containerID="cri-o://0d3d3e6ff01b78a931663c6251387287897e9e19e2021670230cbfb50b83e32c" gracePeriod=10 Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.571773 4935 generic.go:334] "Generic (PLEG): container finished" podID="35e433a5-2009-4c4e-9e85-49810811b71c" containerID="0d3d3e6ff01b78a931663c6251387287897e9e19e2021670230cbfb50b83e32c" exitCode=0 Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.572292 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" event={"ID":"35e433a5-2009-4c4e-9e85-49810811b71c","Type":"ContainerDied","Data":"0d3d3e6ff01b78a931663c6251387287897e9e19e2021670230cbfb50b83e32c"} Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.786019 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.944082 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-config\") pod \"35e433a5-2009-4c4e-9e85-49810811b71c\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.944140 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-sb\") pod \"35e433a5-2009-4c4e-9e85-49810811b71c\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.944166 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-dns-svc\") pod \"35e433a5-2009-4c4e-9e85-49810811b71c\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.944320 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nkjn\" (UniqueName: \"kubernetes.io/projected/35e433a5-2009-4c4e-9e85-49810811b71c-kube-api-access-5nkjn\") pod \"35e433a5-2009-4c4e-9e85-49810811b71c\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.944395 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-openstack-cell1\") pod \"35e433a5-2009-4c4e-9e85-49810811b71c\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.944432 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-nb\") pod \"35e433a5-2009-4c4e-9e85-49810811b71c\" (UID: \"35e433a5-2009-4c4e-9e85-49810811b71c\") " Oct 05 09:03:16 crc kubenswrapper[4935]: I1005 09:03:16.955318 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e433a5-2009-4c4e-9e85-49810811b71c-kube-api-access-5nkjn" (OuterVolumeSpecName: "kube-api-access-5nkjn") pod "35e433a5-2009-4c4e-9e85-49810811b71c" (UID: "35e433a5-2009-4c4e-9e85-49810811b71c"). InnerVolumeSpecName "kube-api-access-5nkjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.010705 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35e433a5-2009-4c4e-9e85-49810811b71c" (UID: "35e433a5-2009-4c4e-9e85-49810811b71c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.018405 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35e433a5-2009-4c4e-9e85-49810811b71c" (UID: "35e433a5-2009-4c4e-9e85-49810811b71c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.027310 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35e433a5-2009-4c4e-9e85-49810811b71c" (UID: "35e433a5-2009-4c4e-9e85-49810811b71c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.041090 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "35e433a5-2009-4c4e-9e85-49810811b71c" (UID: "35e433a5-2009-4c4e-9e85-49810811b71c"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.046838 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.047048 4935 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.047125 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nkjn\" (UniqueName: \"kubernetes.io/projected/35e433a5-2009-4c4e-9e85-49810811b71c-kube-api-access-5nkjn\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.047178 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.047230 4935 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.052187 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-config" (OuterVolumeSpecName: "config") pod "35e433a5-2009-4c4e-9e85-49810811b71c" (UID: "35e433a5-2009-4c4e-9e85-49810811b71c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.148760 4935 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35e433a5-2009-4c4e-9e85-49810811b71c-config\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.583265 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" event={"ID":"35e433a5-2009-4c4e-9e85-49810811b71c","Type":"ContainerDied","Data":"4f4791727142037fad5474831d0e480bacf91095dd773318792166a9ec857934"} Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.583312 4935 scope.go:117] "RemoveContainer" containerID="0d3d3e6ff01b78a931663c6251387287897e9e19e2021670230cbfb50b83e32c" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.583337 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7847dbb94c-bjcc6" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.606454 4935 scope.go:117] "RemoveContainer" containerID="f1e7be0feb6a4890002374f8519d843d4b2988ffe5740599f57bc7aadc12d707" Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.616821 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7847dbb94c-bjcc6"] Oct 05 09:03:17 crc kubenswrapper[4935]: I1005 09:03:17.626696 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7847dbb94c-bjcc6"] Oct 05 09:03:18 crc kubenswrapper[4935]: I1005 09:03:18.790396 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e433a5-2009-4c4e-9e85-49810811b71c" path="/var/lib/kubelet/pods/35e433a5-2009-4c4e-9e85-49810811b71c/volumes" Oct 05 09:03:20 crc kubenswrapper[4935]: I1005 09:03:20.776870 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:03:20 crc kubenswrapper[4935]: E1005 09:03:20.777507 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.933499 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4"] Oct 05 09:03:26 crc kubenswrapper[4935]: E1005 09:03:26.935074 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e433a5-2009-4c4e-9e85-49810811b71c" containerName="dnsmasq-dns" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.935159 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e433a5-2009-4c4e-9e85-49810811b71c" containerName="dnsmasq-dns" Oct 05 09:03:26 crc kubenswrapper[4935]: E1005 09:03:26.935234 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5874184f-b246-4c70-a084-54cb58364d99" containerName="dnsmasq-dns" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.935304 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5874184f-b246-4c70-a084-54cb58364d99" containerName="dnsmasq-dns" Oct 05 09:03:26 crc kubenswrapper[4935]: E1005 09:03:26.935386 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5874184f-b246-4c70-a084-54cb58364d99" containerName="init" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.935440 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5874184f-b246-4c70-a084-54cb58364d99" containerName="init" Oct 05 09:03:26 crc kubenswrapper[4935]: E1005 09:03:26.935494 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e433a5-2009-4c4e-9e85-49810811b71c" containerName="init" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.935547 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e433a5-2009-4c4e-9e85-49810811b71c" containerName="init" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.935822 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5874184f-b246-4c70-a084-54cb58364d99" containerName="dnsmasq-dns" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.935902 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e433a5-2009-4c4e-9e85-49810811b71c" containerName="dnsmasq-dns" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.936690 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.938945 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.939116 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.939326 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.939492 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.952707 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb"] Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.954432 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.956323 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.956474 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.967857 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4"] Oct 05 09:03:26 crc kubenswrapper[4935]: I1005 09:03:26.976291 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb"] Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.069668 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-487h2\" (UniqueName: \"kubernetes.io/projected/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-kube-api-access-487h2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.070753 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.070881 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.070933 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fnh6\" (UniqueName: \"kubernetes.io/projected/a0dab6e1-6403-4221-820a-eb5397c8f884-kube-api-access-4fnh6\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.071539 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.071609 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.071673 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.071742 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.071949 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.173692 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.174001 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-487h2\" (UniqueName: \"kubernetes.io/projected/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-kube-api-access-487h2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.174394 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.174780 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.174909 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fnh6\" (UniqueName: \"kubernetes.io/projected/a0dab6e1-6403-4221-820a-eb5397c8f884-kube-api-access-4fnh6\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.175063 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.175178 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.175306 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.175440 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.180380 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.180420 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.181542 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.181731 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.183599 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.184858 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.186112 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.192650 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fnh6\" (UniqueName: \"kubernetes.io/projected/a0dab6e1-6403-4221-820a-eb5397c8f884-kube-api-access-4fnh6\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.205201 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-487h2\" (UniqueName: \"kubernetes.io/projected/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-kube-api-access-487h2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.254860 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.275194 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.864798 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4"] Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.865347 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:03:27 crc kubenswrapper[4935]: I1005 09:03:27.955212 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb"] Oct 05 09:03:28 crc kubenswrapper[4935]: W1005 09:03:28.017469 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0dab6e1_6403_4221_820a_eb5397c8f884.slice/crio-4dc289c3145148564132bc21b6b38809fcd7ed22d276b27c9f5f73e91b86f356 WatchSource:0}: Error finding container 4dc289c3145148564132bc21b6b38809fcd7ed22d276b27c9f5f73e91b86f356: Status 404 returned error can't find the container with id 4dc289c3145148564132bc21b6b38809fcd7ed22d276b27c9f5f73e91b86f356 Oct 05 09:03:28 crc kubenswrapper[4935]: I1005 09:03:28.715727 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" event={"ID":"a0dab6e1-6403-4221-820a-eb5397c8f884","Type":"ContainerStarted","Data":"4dc289c3145148564132bc21b6b38809fcd7ed22d276b27c9f5f73e91b86f356"} Oct 05 09:03:28 crc kubenswrapper[4935]: I1005 09:03:28.718282 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" event={"ID":"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da","Type":"ContainerStarted","Data":"ee7262a029bb44635755e65f2db9ed4fc02090386accabe5292e599754504647"} Oct 05 09:03:34 crc kubenswrapper[4935]: I1005 09:03:34.777198 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:03:34 crc kubenswrapper[4935]: E1005 09:03:34.778071 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:03:37 crc kubenswrapper[4935]: I1005 09:03:37.871274 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" event={"ID":"a0dab6e1-6403-4221-820a-eb5397c8f884","Type":"ContainerStarted","Data":"bfea6cbe9c2057824920fdc03d9ff5a2756f0bdd973f6aeefc3c3ee9aa840591"} Oct 05 09:03:37 crc kubenswrapper[4935]: I1005 09:03:37.876878 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" event={"ID":"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da","Type":"ContainerStarted","Data":"8995665617ecb16fcaa4467bca612bb7ca622d9e4d457f5d8f820e9126bce8f7"} Oct 05 09:03:37 crc kubenswrapper[4935]: I1005 09:03:37.893702 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" podStartSLOduration=2.401440164 podStartE2EDuration="11.893681909s" podCreationTimestamp="2025-10-05 09:03:26 +0000 UTC" firstStartedPulling="2025-10-05 09:03:28.019761434 +0000 UTC m=+7841.902387894" lastFinishedPulling="2025-10-05 09:03:37.512003179 +0000 UTC m=+7851.394629639" observedRunningTime="2025-10-05 09:03:37.887796743 +0000 UTC m=+7851.770423203" watchObservedRunningTime="2025-10-05 09:03:37.893681909 +0000 UTC m=+7851.776308369" Oct 05 09:03:37 crc kubenswrapper[4935]: I1005 09:03:37.913570 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" podStartSLOduration=2.295075013 podStartE2EDuration="11.913554055s" podCreationTimestamp="2025-10-05 09:03:26 +0000 UTC" firstStartedPulling="2025-10-05 09:03:27.865154675 +0000 UTC m=+7841.747781135" lastFinishedPulling="2025-10-05 09:03:37.483633677 +0000 UTC m=+7851.366260177" observedRunningTime="2025-10-05 09:03:37.905880862 +0000 UTC m=+7851.788507362" watchObservedRunningTime="2025-10-05 09:03:37.913554055 +0000 UTC m=+7851.796180515" Oct 05 09:03:43 crc kubenswrapper[4935]: I1005 09:03:43.049354 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-lp2bz"] Oct 05 09:03:43 crc kubenswrapper[4935]: I1005 09:03:43.063614 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-lp2bz"] Oct 05 09:03:44 crc kubenswrapper[4935]: I1005 09:03:44.788713 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a680f5ef-08d9-4f31-a4a5-5230086909a6" path="/var/lib/kubelet/pods/a680f5ef-08d9-4f31-a4a5-5230086909a6/volumes" Oct 05 09:03:47 crc kubenswrapper[4935]: I1005 09:03:47.991518 4935 generic.go:334] "Generic (PLEG): container finished" podID="a0dab6e1-6403-4221-820a-eb5397c8f884" containerID="bfea6cbe9c2057824920fdc03d9ff5a2756f0bdd973f6aeefc3c3ee9aa840591" exitCode=0 Oct 05 09:03:47 crc kubenswrapper[4935]: I1005 09:03:47.991614 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" event={"ID":"a0dab6e1-6403-4221-820a-eb5397c8f884","Type":"ContainerDied","Data":"bfea6cbe9c2057824920fdc03d9ff5a2756f0bdd973f6aeefc3c3ee9aa840591"} Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.006299 4935 generic.go:334] "Generic (PLEG): container finished" podID="7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" containerID="8995665617ecb16fcaa4467bca612bb7ca622d9e4d457f5d8f820e9126bce8f7" exitCode=0 Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.006392 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" event={"ID":"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da","Type":"ContainerDied","Data":"8995665617ecb16fcaa4467bca612bb7ca622d9e4d457f5d8f820e9126bce8f7"} Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.579006 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.674928 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-ssh-key\") pod \"a0dab6e1-6403-4221-820a-eb5397c8f884\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.674993 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-pre-adoption-validation-combined-ca-bundle\") pod \"a0dab6e1-6403-4221-820a-eb5397c8f884\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.675069 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-inventory\") pod \"a0dab6e1-6403-4221-820a-eb5397c8f884\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.675226 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fnh6\" (UniqueName: \"kubernetes.io/projected/a0dab6e1-6403-4221-820a-eb5397c8f884-kube-api-access-4fnh6\") pod \"a0dab6e1-6403-4221-820a-eb5397c8f884\" (UID: \"a0dab6e1-6403-4221-820a-eb5397c8f884\") " Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.681742 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "a0dab6e1-6403-4221-820a-eb5397c8f884" (UID: "a0dab6e1-6403-4221-820a-eb5397c8f884"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.695754 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0dab6e1-6403-4221-820a-eb5397c8f884-kube-api-access-4fnh6" (OuterVolumeSpecName: "kube-api-access-4fnh6") pod "a0dab6e1-6403-4221-820a-eb5397c8f884" (UID: "a0dab6e1-6403-4221-820a-eb5397c8f884"). InnerVolumeSpecName "kube-api-access-4fnh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.718286 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a0dab6e1-6403-4221-820a-eb5397c8f884" (UID: "a0dab6e1-6403-4221-820a-eb5397c8f884"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.730303 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-inventory" (OuterVolumeSpecName: "inventory") pod "a0dab6e1-6403-4221-820a-eb5397c8f884" (UID: "a0dab6e1-6403-4221-820a-eb5397c8f884"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.777552 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:03:49 crc kubenswrapper[4935]: E1005 09:03:49.778348 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.779828 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fnh6\" (UniqueName: \"kubernetes.io/projected/a0dab6e1-6403-4221-820a-eb5397c8f884-kube-api-access-4fnh6\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.779871 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.779911 4935 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:49 crc kubenswrapper[4935]: I1005 09:03:49.779932 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0dab6e1-6403-4221-820a-eb5397c8f884-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.022240 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.022235 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb" event={"ID":"a0dab6e1-6403-4221-820a-eb5397c8f884","Type":"ContainerDied","Data":"4dc289c3145148564132bc21b6b38809fcd7ed22d276b27c9f5f73e91b86f356"} Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.022831 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dc289c3145148564132bc21b6b38809fcd7ed22d276b27c9f5f73e91b86f356" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.467701 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.596276 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-487h2\" (UniqueName: \"kubernetes.io/projected/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-kube-api-access-487h2\") pod \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.596381 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ssh-key\") pod \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.596454 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-inventory\") pod \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.596493 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ceph\") pod \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.596596 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-pre-adoption-validation-combined-ca-bundle\") pod \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\" (UID: \"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da\") " Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.603168 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ceph" (OuterVolumeSpecName: "ceph") pod "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" (UID: "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.603225 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-kube-api-access-487h2" (OuterVolumeSpecName: "kube-api-access-487h2") pod "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" (UID: "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da"). InnerVolumeSpecName "kube-api-access-487h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.603790 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" (UID: "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.644649 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-inventory" (OuterVolumeSpecName: "inventory") pod "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" (UID: "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.652093 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" (UID: "7e2ad6c2-e6d7-4592-ab04-0b13b92be5da"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.700400 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.700646 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.700732 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.700823 4935 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:50 crc kubenswrapper[4935]: I1005 09:03:50.700937 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-487h2\" (UniqueName: \"kubernetes.io/projected/7e2ad6c2-e6d7-4592-ab04-0b13b92be5da-kube-api-access-487h2\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:51 crc kubenswrapper[4935]: I1005 09:03:51.036213 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" event={"ID":"7e2ad6c2-e6d7-4592-ab04-0b13b92be5da","Type":"ContainerDied","Data":"ee7262a029bb44635755e65f2db9ed4fc02090386accabe5292e599754504647"} Oct 05 09:03:51 crc kubenswrapper[4935]: I1005 09:03:51.036274 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee7262a029bb44635755e65f2db9ed4fc02090386accabe5292e599754504647" Oct 05 09:03:51 crc kubenswrapper[4935]: I1005 09:03:51.036350 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4" Oct 05 09:03:53 crc kubenswrapper[4935]: I1005 09:03:53.031788 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-fd41-account-create-nkmgw"] Oct 05 09:03:53 crc kubenswrapper[4935]: I1005 09:03:53.041862 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-fd41-account-create-nkmgw"] Oct 05 09:03:54 crc kubenswrapper[4935]: I1005 09:03:54.801960 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="313db7f6-e689-4071-87d8-71ea5bc7cb30" path="/var/lib/kubelet/pods/313db7f6-e689-4071-87d8-71ea5bc7cb30/volumes" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.177963 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx"] Oct 05 09:04:00 crc kubenswrapper[4935]: E1005 09:04:00.179016 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0dab6e1-6403-4221-820a-eb5397c8f884" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.179035 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0dab6e1-6403-4221-820a-eb5397c8f884" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 05 09:04:00 crc kubenswrapper[4935]: E1005 09:04:00.179075 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.179083 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.179307 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0dab6e1-6403-4221-820a-eb5397c8f884" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.179345 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e2ad6c2-e6d7-4592-ab04-0b13b92be5da" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.180165 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.183332 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.183752 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.184129 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.184304 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.195710 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp"] Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.197004 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.199731 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.200093 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.206860 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp"] Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.215357 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx"] Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.328930 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr69b\" (UniqueName: \"kubernetes.io/projected/aad90193-5bba-4acd-ac25-d67d21a165d3-kube-api-access-gr69b\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.329000 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.329031 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.329048 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.329361 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.329533 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.329593 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np8gt\" (UniqueName: \"kubernetes.io/projected/097a7b90-7e36-4aa4-b9df-e17362922f55-kube-api-access-np8gt\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.329668 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.329747 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432076 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432124 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432191 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432230 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432251 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np8gt\" (UniqueName: \"kubernetes.io/projected/097a7b90-7e36-4aa4-b9df-e17362922f55-kube-api-access-np8gt\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432285 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432343 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432386 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr69b\" (UniqueName: \"kubernetes.io/projected/aad90193-5bba-4acd-ac25-d67d21a165d3-kube-api-access-gr69b\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.432423 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.438789 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.439081 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.444800 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.445033 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.445130 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.448704 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.449583 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr69b\" (UniqueName: \"kubernetes.io/projected/aad90193-5bba-4acd-ac25-d67d21a165d3-kube-api-access-gr69b\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.450633 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.461249 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np8gt\" (UniqueName: \"kubernetes.io/projected/097a7b90-7e36-4aa4-b9df-e17362922f55-kube-api-access-np8gt\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.547839 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:04:00 crc kubenswrapper[4935]: I1005 09:04:00.559187 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:04:01 crc kubenswrapper[4935]: I1005 09:04:01.146317 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp"] Oct 05 09:04:01 crc kubenswrapper[4935]: I1005 09:04:01.162332 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" event={"ID":"aad90193-5bba-4acd-ac25-d67d21a165d3","Type":"ContainerStarted","Data":"db452dc4ac9d3592502936c9fa30cf4d5cbd57ad0b345c10364d913e01d6737b"} Oct 05 09:04:01 crc kubenswrapper[4935]: I1005 09:04:01.245636 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx"] Oct 05 09:04:01 crc kubenswrapper[4935]: W1005 09:04:01.249625 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod097a7b90_7e36_4aa4_b9df_e17362922f55.slice/crio-707e621d1347ca891aed52e46b35671344ca7c6c503380cf0f4c188a56322a09 WatchSource:0}: Error finding container 707e621d1347ca891aed52e46b35671344ca7c6c503380cf0f4c188a56322a09: Status 404 returned error can't find the container with id 707e621d1347ca891aed52e46b35671344ca7c6c503380cf0f4c188a56322a09 Oct 05 09:04:01 crc kubenswrapper[4935]: I1005 09:04:01.777052 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:04:01 crc kubenswrapper[4935]: E1005 09:04:01.777462 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:04:02 crc kubenswrapper[4935]: I1005 09:04:02.047463 4935 scope.go:117] "RemoveContainer" containerID="128166db61018ffddcb331e01c51278b845e221547a789749ca241d6a183dd99" Oct 05 09:04:02 crc kubenswrapper[4935]: I1005 09:04:02.089210 4935 scope.go:117] "RemoveContainer" containerID="360fa3306cc49a1bdddcc86ff46c7160a777c9ee86c27f808dc4b62093559767" Oct 05 09:04:02 crc kubenswrapper[4935]: I1005 09:04:02.177004 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" event={"ID":"aad90193-5bba-4acd-ac25-d67d21a165d3","Type":"ContainerStarted","Data":"aa2f79bf9d1bccdfa51f6b71201d642193908f75bb5862355fdade19b12fc868"} Oct 05 09:04:02 crc kubenswrapper[4935]: I1005 09:04:02.193301 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" event={"ID":"097a7b90-7e36-4aa4-b9df-e17362922f55","Type":"ContainerStarted","Data":"2035fcb4cf084fe243d6116a7c4b54048adb85ac94aad7a62adcaa6a9eb0f0a9"} Oct 05 09:04:02 crc kubenswrapper[4935]: I1005 09:04:02.193357 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" event={"ID":"097a7b90-7e36-4aa4-b9df-e17362922f55","Type":"ContainerStarted","Data":"707e621d1347ca891aed52e46b35671344ca7c6c503380cf0f4c188a56322a09"} Oct 05 09:04:02 crc kubenswrapper[4935]: I1005 09:04:02.196338 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" podStartSLOduration=1.6720374850000002 podStartE2EDuration="2.196322436s" podCreationTimestamp="2025-10-05 09:04:00 +0000 UTC" firstStartedPulling="2025-10-05 09:04:01.152010048 +0000 UTC m=+7875.034636508" lastFinishedPulling="2025-10-05 09:04:01.676294959 +0000 UTC m=+7875.558921459" observedRunningTime="2025-10-05 09:04:02.19194852 +0000 UTC m=+7876.074574980" watchObservedRunningTime="2025-10-05 09:04:02.196322436 +0000 UTC m=+7876.078948896" Oct 05 09:04:02 crc kubenswrapper[4935]: I1005 09:04:02.221004 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" podStartSLOduration=1.774390279 podStartE2EDuration="2.22098626s" podCreationTimestamp="2025-10-05 09:04:00 +0000 UTC" firstStartedPulling="2025-10-05 09:04:01.251762513 +0000 UTC m=+7875.134388973" lastFinishedPulling="2025-10-05 09:04:01.698358454 +0000 UTC m=+7875.580984954" observedRunningTime="2025-10-05 09:04:02.211290593 +0000 UTC m=+7876.093917053" watchObservedRunningTime="2025-10-05 09:04:02.22098626 +0000 UTC m=+7876.103612710" Oct 05 09:04:13 crc kubenswrapper[4935]: I1005 09:04:13.778673 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:04:13 crc kubenswrapper[4935]: E1005 09:04:13.779761 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:04:20 crc kubenswrapper[4935]: I1005 09:04:20.045119 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-kjp9m"] Oct 05 09:04:20 crc kubenswrapper[4935]: I1005 09:04:20.052927 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-kjp9m"] Oct 05 09:04:20 crc kubenswrapper[4935]: I1005 09:04:20.790694 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="850ca766-4edf-4518-9e91-bf08ce1d119e" path="/var/lib/kubelet/pods/850ca766-4edf-4518-9e91-bf08ce1d119e/volumes" Oct 05 09:04:27 crc kubenswrapper[4935]: I1005 09:04:27.777615 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:04:27 crc kubenswrapper[4935]: E1005 09:04:27.778953 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:04:37 crc kubenswrapper[4935]: I1005 09:04:37.902361 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5cjpb"] Oct 05 09:04:37 crc kubenswrapper[4935]: I1005 09:04:37.905487 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:37 crc kubenswrapper[4935]: I1005 09:04:37.930228 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5cjpb"] Oct 05 09:04:37 crc kubenswrapper[4935]: I1005 09:04:37.986796 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-catalog-content\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:37 crc kubenswrapper[4935]: I1005 09:04:37.986852 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-utilities\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:37 crc kubenswrapper[4935]: I1005 09:04:37.987022 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c49p\" (UniqueName: \"kubernetes.io/projected/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-kube-api-access-9c49p\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.089389 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c49p\" (UniqueName: \"kubernetes.io/projected/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-kube-api-access-9c49p\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.089528 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-catalog-content\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.089551 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-utilities\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.090018 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-utilities\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.090391 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-catalog-content\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.113762 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c49p\" (UniqueName: \"kubernetes.io/projected/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-kube-api-access-9c49p\") pod \"redhat-operators-5cjpb\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.246013 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.729173 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5cjpb"] Oct 05 09:04:38 crc kubenswrapper[4935]: I1005 09:04:38.777913 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:04:38 crc kubenswrapper[4935]: E1005 09:04:38.778198 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:04:39 crc kubenswrapper[4935]: I1005 09:04:39.653903 4935 generic.go:334] "Generic (PLEG): container finished" podID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerID="22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6" exitCode=0 Oct 05 09:04:39 crc kubenswrapper[4935]: I1005 09:04:39.654012 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cjpb" event={"ID":"78831ce6-216f-46aa-8f3a-dc629b4ccbf1","Type":"ContainerDied","Data":"22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6"} Oct 05 09:04:39 crc kubenswrapper[4935]: I1005 09:04:39.654240 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cjpb" event={"ID":"78831ce6-216f-46aa-8f3a-dc629b4ccbf1","Type":"ContainerStarted","Data":"8b3e4ff895daa9e3b0f715abcf9d45893d8d3aef578df189bdcbe03b0f2bf529"} Oct 05 09:04:40 crc kubenswrapper[4935]: I1005 09:04:40.664226 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cjpb" event={"ID":"78831ce6-216f-46aa-8f3a-dc629b4ccbf1","Type":"ContainerStarted","Data":"929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19"} Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.098412 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rjcqz"] Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.101843 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.109418 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rjcqz"] Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.262077 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hnr6\" (UniqueName: \"kubernetes.io/projected/4a9a4be3-ebdb-459f-a35f-089cfaac3920-kube-api-access-2hnr6\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.262239 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-catalog-content\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.262301 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-utilities\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.363836 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-catalog-content\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.363941 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-utilities\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.363989 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hnr6\" (UniqueName: \"kubernetes.io/projected/4a9a4be3-ebdb-459f-a35f-089cfaac3920-kube-api-access-2hnr6\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.364448 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-catalog-content\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.364591 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-utilities\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.384300 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hnr6\" (UniqueName: \"kubernetes.io/projected/4a9a4be3-ebdb-459f-a35f-089cfaac3920-kube-api-access-2hnr6\") pod \"certified-operators-rjcqz\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.432236 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:41 crc kubenswrapper[4935]: I1005 09:04:41.946756 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rjcqz"] Oct 05 09:04:42 crc kubenswrapper[4935]: I1005 09:04:42.700274 4935 generic.go:334] "Generic (PLEG): container finished" podID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerID="d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f" exitCode=0 Oct 05 09:04:42 crc kubenswrapper[4935]: I1005 09:04:42.700716 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjcqz" event={"ID":"4a9a4be3-ebdb-459f-a35f-089cfaac3920","Type":"ContainerDied","Data":"d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f"} Oct 05 09:04:42 crc kubenswrapper[4935]: I1005 09:04:42.700760 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjcqz" event={"ID":"4a9a4be3-ebdb-459f-a35f-089cfaac3920","Type":"ContainerStarted","Data":"c7f7887b1a4dfd2bcf8198f5e29313f52d63fc6cb3170cda1b895a4f93cd4a8e"} Oct 05 09:04:44 crc kubenswrapper[4935]: I1005 09:04:44.728574 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjcqz" event={"ID":"4a9a4be3-ebdb-459f-a35f-089cfaac3920","Type":"ContainerStarted","Data":"07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f"} Oct 05 09:04:45 crc kubenswrapper[4935]: I1005 09:04:45.741586 4935 generic.go:334] "Generic (PLEG): container finished" podID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerID="929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19" exitCode=0 Oct 05 09:04:45 crc kubenswrapper[4935]: I1005 09:04:45.741672 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cjpb" event={"ID":"78831ce6-216f-46aa-8f3a-dc629b4ccbf1","Type":"ContainerDied","Data":"929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19"} Oct 05 09:04:46 crc kubenswrapper[4935]: I1005 09:04:46.754330 4935 generic.go:334] "Generic (PLEG): container finished" podID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerID="07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f" exitCode=0 Oct 05 09:04:46 crc kubenswrapper[4935]: I1005 09:04:46.754548 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjcqz" event={"ID":"4a9a4be3-ebdb-459f-a35f-089cfaac3920","Type":"ContainerDied","Data":"07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f"} Oct 05 09:04:46 crc kubenswrapper[4935]: I1005 09:04:46.758081 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cjpb" event={"ID":"78831ce6-216f-46aa-8f3a-dc629b4ccbf1","Type":"ContainerStarted","Data":"adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805"} Oct 05 09:04:46 crc kubenswrapper[4935]: I1005 09:04:46.800311 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5cjpb" podStartSLOduration=3.296583567 podStartE2EDuration="9.800293747s" podCreationTimestamp="2025-10-05 09:04:37 +0000 UTC" firstStartedPulling="2025-10-05 09:04:39.656886577 +0000 UTC m=+7913.539513077" lastFinishedPulling="2025-10-05 09:04:46.160596797 +0000 UTC m=+7920.043223257" observedRunningTime="2025-10-05 09:04:46.796768274 +0000 UTC m=+7920.679394734" watchObservedRunningTime="2025-10-05 09:04:46.800293747 +0000 UTC m=+7920.682920207" Oct 05 09:04:48 crc kubenswrapper[4935]: I1005 09:04:48.246941 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:48 crc kubenswrapper[4935]: I1005 09:04:48.247551 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:04:48 crc kubenswrapper[4935]: I1005 09:04:48.788469 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjcqz" event={"ID":"4a9a4be3-ebdb-459f-a35f-089cfaac3920","Type":"ContainerStarted","Data":"7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f"} Oct 05 09:04:48 crc kubenswrapper[4935]: I1005 09:04:48.808955 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rjcqz" podStartSLOduration=2.446757406 podStartE2EDuration="7.808938771s" podCreationTimestamp="2025-10-05 09:04:41 +0000 UTC" firstStartedPulling="2025-10-05 09:04:42.703089899 +0000 UTC m=+7916.585716369" lastFinishedPulling="2025-10-05 09:04:48.065271274 +0000 UTC m=+7921.947897734" observedRunningTime="2025-10-05 09:04:48.806586669 +0000 UTC m=+7922.689213129" watchObservedRunningTime="2025-10-05 09:04:48.808938771 +0000 UTC m=+7922.691565231" Oct 05 09:04:49 crc kubenswrapper[4935]: I1005 09:04:49.301728 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5cjpb" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="registry-server" probeResult="failure" output=< Oct 05 09:04:49 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 09:04:49 crc kubenswrapper[4935]: > Oct 05 09:04:51 crc kubenswrapper[4935]: I1005 09:04:51.432671 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:51 crc kubenswrapper[4935]: I1005 09:04:51.433876 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:51 crc kubenswrapper[4935]: I1005 09:04:51.483224 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:04:51 crc kubenswrapper[4935]: I1005 09:04:51.777319 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:04:51 crc kubenswrapper[4935]: E1005 09:04:51.777634 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:04:59 crc kubenswrapper[4935]: I1005 09:04:59.315940 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5cjpb" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="registry-server" probeResult="failure" output=< Oct 05 09:04:59 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 09:04:59 crc kubenswrapper[4935]: > Oct 05 09:05:01 crc kubenswrapper[4935]: I1005 09:05:01.557400 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:05:01 crc kubenswrapper[4935]: I1005 09:05:01.602431 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rjcqz"] Oct 05 09:05:01 crc kubenswrapper[4935]: I1005 09:05:01.890970 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rjcqz" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerName="registry-server" containerID="cri-o://7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f" gracePeriod=2 Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.247780 4935 scope.go:117] "RemoveContainer" containerID="d63098e78c0bb39eebbde224ac01bbefa97d873c7fda7d06b3a6e26adb2978d6" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.468100 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.624854 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-catalog-content\") pod \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.625023 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-utilities\") pod \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.625269 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hnr6\" (UniqueName: \"kubernetes.io/projected/4a9a4be3-ebdb-459f-a35f-089cfaac3920-kube-api-access-2hnr6\") pod \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\" (UID: \"4a9a4be3-ebdb-459f-a35f-089cfaac3920\") " Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.625635 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-utilities" (OuterVolumeSpecName: "utilities") pod "4a9a4be3-ebdb-459f-a35f-089cfaac3920" (UID: "4a9a4be3-ebdb-459f-a35f-089cfaac3920"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.626109 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.631320 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a9a4be3-ebdb-459f-a35f-089cfaac3920-kube-api-access-2hnr6" (OuterVolumeSpecName: "kube-api-access-2hnr6") pod "4a9a4be3-ebdb-459f-a35f-089cfaac3920" (UID: "4a9a4be3-ebdb-459f-a35f-089cfaac3920"). InnerVolumeSpecName "kube-api-access-2hnr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.667256 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a9a4be3-ebdb-459f-a35f-089cfaac3920" (UID: "4a9a4be3-ebdb-459f-a35f-089cfaac3920"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.727797 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hnr6\" (UniqueName: \"kubernetes.io/projected/4a9a4be3-ebdb-459f-a35f-089cfaac3920-kube-api-access-2hnr6\") on node \"crc\" DevicePath \"\"" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.727839 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9a4be3-ebdb-459f-a35f-089cfaac3920-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.900742 4935 generic.go:334] "Generic (PLEG): container finished" podID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerID="7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f" exitCode=0 Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.900812 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjcqz" event={"ID":"4a9a4be3-ebdb-459f-a35f-089cfaac3920","Type":"ContainerDied","Data":"7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f"} Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.900839 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rjcqz" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.900865 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjcqz" event={"ID":"4a9a4be3-ebdb-459f-a35f-089cfaac3920","Type":"ContainerDied","Data":"c7f7887b1a4dfd2bcf8198f5e29313f52d63fc6cb3170cda1b895a4f93cd4a8e"} Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.900884 4935 scope.go:117] "RemoveContainer" containerID="7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.925554 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rjcqz"] Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.932715 4935 scope.go:117] "RemoveContainer" containerID="07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.934488 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rjcqz"] Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.956209 4935 scope.go:117] "RemoveContainer" containerID="d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.986713 4935 scope.go:117] "RemoveContainer" containerID="7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f" Oct 05 09:05:02 crc kubenswrapper[4935]: E1005 09:05:02.987289 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f\": container with ID starting with 7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f not found: ID does not exist" containerID="7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.987322 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f"} err="failed to get container status \"7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f\": rpc error: code = NotFound desc = could not find container \"7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f\": container with ID starting with 7f7ec46c2a29b192d23c8c6dc79a55fa0e9bdf941adbc639b3440936f5adf69f not found: ID does not exist" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.987346 4935 scope.go:117] "RemoveContainer" containerID="07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f" Oct 05 09:05:02 crc kubenswrapper[4935]: E1005 09:05:02.987620 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f\": container with ID starting with 07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f not found: ID does not exist" containerID="07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.987641 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f"} err="failed to get container status \"07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f\": rpc error: code = NotFound desc = could not find container \"07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f\": container with ID starting with 07cb7447aaae0147fc4dfe68863fa4aeddc472003d11de7a2a191085c2d91a9f not found: ID does not exist" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.987654 4935 scope.go:117] "RemoveContainer" containerID="d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f" Oct 05 09:05:02 crc kubenswrapper[4935]: E1005 09:05:02.987839 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f\": container with ID starting with d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f not found: ID does not exist" containerID="d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f" Oct 05 09:05:02 crc kubenswrapper[4935]: I1005 09:05:02.987858 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f"} err="failed to get container status \"d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f\": rpc error: code = NotFound desc = could not find container \"d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f\": container with ID starting with d439b275c7f1328f6ce88e42fa3780d37d5b1d540fd16a4090f916346a74c68f not found: ID does not exist" Oct 05 09:05:03 crc kubenswrapper[4935]: I1005 09:05:03.778231 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:05:03 crc kubenswrapper[4935]: E1005 09:05:03.778736 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:05:04 crc kubenswrapper[4935]: I1005 09:05:04.789397 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" path="/var/lib/kubelet/pods/4a9a4be3-ebdb-459f-a35f-089cfaac3920/volumes" Oct 05 09:05:09 crc kubenswrapper[4935]: I1005 09:05:09.297228 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5cjpb" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="registry-server" probeResult="failure" output=< Oct 05 09:05:09 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 09:05:09 crc kubenswrapper[4935]: > Oct 05 09:05:14 crc kubenswrapper[4935]: I1005 09:05:14.777051 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:05:14 crc kubenswrapper[4935]: E1005 09:05:14.777797 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:05:18 crc kubenswrapper[4935]: I1005 09:05:18.299354 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:05:18 crc kubenswrapper[4935]: I1005 09:05:18.373001 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:05:18 crc kubenswrapper[4935]: I1005 09:05:18.547678 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5cjpb"] Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.070796 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5cjpb" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="registry-server" containerID="cri-o://adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805" gracePeriod=2 Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.632007 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.736918 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-utilities\") pod \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.737105 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c49p\" (UniqueName: \"kubernetes.io/projected/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-kube-api-access-9c49p\") pod \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.737235 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-catalog-content\") pod \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\" (UID: \"78831ce6-216f-46aa-8f3a-dc629b4ccbf1\") " Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.737871 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-utilities" (OuterVolumeSpecName: "utilities") pod "78831ce6-216f-46aa-8f3a-dc629b4ccbf1" (UID: "78831ce6-216f-46aa-8f3a-dc629b4ccbf1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.746376 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-kube-api-access-9c49p" (OuterVolumeSpecName: "kube-api-access-9c49p") pod "78831ce6-216f-46aa-8f3a-dc629b4ccbf1" (UID: "78831ce6-216f-46aa-8f3a-dc629b4ccbf1"). InnerVolumeSpecName "kube-api-access-9c49p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.819640 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78831ce6-216f-46aa-8f3a-dc629b4ccbf1" (UID: "78831ce6-216f-46aa-8f3a-dc629b4ccbf1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.839597 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c49p\" (UniqueName: \"kubernetes.io/projected/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-kube-api-access-9c49p\") on node \"crc\" DevicePath \"\"" Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.839634 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:05:20 crc kubenswrapper[4935]: I1005 09:05:20.839646 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78831ce6-216f-46aa-8f3a-dc629b4ccbf1-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.082676 4935 generic.go:334] "Generic (PLEG): container finished" podID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerID="adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805" exitCode=0 Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.082717 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cjpb" event={"ID":"78831ce6-216f-46aa-8f3a-dc629b4ccbf1","Type":"ContainerDied","Data":"adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805"} Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.082752 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cjpb" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.082770 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cjpb" event={"ID":"78831ce6-216f-46aa-8f3a-dc629b4ccbf1","Type":"ContainerDied","Data":"8b3e4ff895daa9e3b0f715abcf9d45893d8d3aef578df189bdcbe03b0f2bf529"} Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.082793 4935 scope.go:117] "RemoveContainer" containerID="adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.129619 4935 scope.go:117] "RemoveContainer" containerID="929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.132662 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5cjpb"] Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.143494 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5cjpb"] Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.156012 4935 scope.go:117] "RemoveContainer" containerID="22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.209981 4935 scope.go:117] "RemoveContainer" containerID="adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805" Oct 05 09:05:21 crc kubenswrapper[4935]: E1005 09:05:21.210751 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805\": container with ID starting with adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805 not found: ID does not exist" containerID="adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.210793 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805"} err="failed to get container status \"adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805\": rpc error: code = NotFound desc = could not find container \"adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805\": container with ID starting with adf1d86d2dbb75e171aeacb38b64f325180db2c5b2f4fecc62720db5ca424805 not found: ID does not exist" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.210818 4935 scope.go:117] "RemoveContainer" containerID="929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19" Oct 05 09:05:21 crc kubenswrapper[4935]: E1005 09:05:21.211259 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19\": container with ID starting with 929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19 not found: ID does not exist" containerID="929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.211312 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19"} err="failed to get container status \"929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19\": rpc error: code = NotFound desc = could not find container \"929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19\": container with ID starting with 929f6b5aaa89a1bc99a0b5ccba630209b8c6dd0d0d77a076b3a0d6098e37ff19 not found: ID does not exist" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.211344 4935 scope.go:117] "RemoveContainer" containerID="22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6" Oct 05 09:05:21 crc kubenswrapper[4935]: E1005 09:05:21.211698 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6\": container with ID starting with 22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6 not found: ID does not exist" containerID="22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6" Oct 05 09:05:21 crc kubenswrapper[4935]: I1005 09:05:21.211722 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6"} err="failed to get container status \"22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6\": rpc error: code = NotFound desc = could not find container \"22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6\": container with ID starting with 22cfddc82c46e56053c74be3b62c0f5aceb44a4108fce85ff371270fc582aab6 not found: ID does not exist" Oct 05 09:05:22 crc kubenswrapper[4935]: I1005 09:05:22.792732 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" path="/var/lib/kubelet/pods/78831ce6-216f-46aa-8f3a-dc629b4ccbf1/volumes" Oct 05 09:05:27 crc kubenswrapper[4935]: I1005 09:05:27.776812 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:05:27 crc kubenswrapper[4935]: E1005 09:05:27.777453 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:05:40 crc kubenswrapper[4935]: I1005 09:05:40.778386 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:05:40 crc kubenswrapper[4935]: E1005 09:05:40.779789 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:05:54 crc kubenswrapper[4935]: I1005 09:05:54.776808 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:05:55 crc kubenswrapper[4935]: I1005 09:05:55.502463 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"b3f5371d7f74bebfc1a9f20169e0444802d8ff98646f3001d261f7f1c9425f2f"} Oct 05 09:08:10 crc kubenswrapper[4935]: I1005 09:08:10.047089 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-m4w8v"] Oct 05 09:08:10 crc kubenswrapper[4935]: I1005 09:08:10.055747 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-m4w8v"] Oct 05 09:08:10 crc kubenswrapper[4935]: I1005 09:08:10.789230 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbd77f57-fc8f-48d7-afcb-37e67611049f" path="/var/lib/kubelet/pods/cbd77f57-fc8f-48d7-afcb-37e67611049f/volumes" Oct 05 09:08:14 crc kubenswrapper[4935]: I1005 09:08:14.289620 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:08:14 crc kubenswrapper[4935]: I1005 09:08:14.290091 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:08:19 crc kubenswrapper[4935]: I1005 09:08:19.032248 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-6d32-account-create-9qkt6"] Oct 05 09:08:19 crc kubenswrapper[4935]: I1005 09:08:19.042360 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-6d32-account-create-9qkt6"] Oct 05 09:08:20 crc kubenswrapper[4935]: I1005 09:08:20.795400 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adf81e84-5745-4f16-acc2-3ee390046548" path="/var/lib/kubelet/pods/adf81e84-5745-4f16-acc2-3ee390046548/volumes" Oct 05 09:08:32 crc kubenswrapper[4935]: I1005 09:08:32.247348 4935 generic.go:334] "Generic (PLEG): container finished" podID="aad90193-5bba-4acd-ac25-d67d21a165d3" containerID="aa2f79bf9d1bccdfa51f6b71201d642193908f75bb5862355fdade19b12fc868" exitCode=0 Oct 05 09:08:32 crc kubenswrapper[4935]: I1005 09:08:32.247495 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" event={"ID":"aad90193-5bba-4acd-ac25-d67d21a165d3","Type":"ContainerDied","Data":"aa2f79bf9d1bccdfa51f6b71201d642193908f75bb5862355fdade19b12fc868"} Oct 05 09:08:33 crc kubenswrapper[4935]: I1005 09:08:33.875712 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:08:33 crc kubenswrapper[4935]: I1005 09:08:33.977423 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr69b\" (UniqueName: \"kubernetes.io/projected/aad90193-5bba-4acd-ac25-d67d21a165d3-kube-api-access-gr69b\") pod \"aad90193-5bba-4acd-ac25-d67d21a165d3\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " Oct 05 09:08:33 crc kubenswrapper[4935]: I1005 09:08:33.977640 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-tripleo-cleanup-combined-ca-bundle\") pod \"aad90193-5bba-4acd-ac25-d67d21a165d3\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " Oct 05 09:08:33 crc kubenswrapper[4935]: I1005 09:08:33.977713 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-inventory\") pod \"aad90193-5bba-4acd-ac25-d67d21a165d3\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " Oct 05 09:08:33 crc kubenswrapper[4935]: I1005 09:08:33.977759 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-ssh-key\") pod \"aad90193-5bba-4acd-ac25-d67d21a165d3\" (UID: \"aad90193-5bba-4acd-ac25-d67d21a165d3\") " Oct 05 09:08:33 crc kubenswrapper[4935]: I1005 09:08:33.983718 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "aad90193-5bba-4acd-ac25-d67d21a165d3" (UID: "aad90193-5bba-4acd-ac25-d67d21a165d3"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:33 crc kubenswrapper[4935]: I1005 09:08:33.984233 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad90193-5bba-4acd-ac25-d67d21a165d3-kube-api-access-gr69b" (OuterVolumeSpecName: "kube-api-access-gr69b") pod "aad90193-5bba-4acd-ac25-d67d21a165d3" (UID: "aad90193-5bba-4acd-ac25-d67d21a165d3"). InnerVolumeSpecName "kube-api-access-gr69b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.035961 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aad90193-5bba-4acd-ac25-d67d21a165d3" (UID: "aad90193-5bba-4acd-ac25-d67d21a165d3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.036034 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-9pd72"] Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.041634 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-inventory" (OuterVolumeSpecName: "inventory") pod "aad90193-5bba-4acd-ac25-d67d21a165d3" (UID: "aad90193-5bba-4acd-ac25-d67d21a165d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.047139 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-9pd72"] Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.079813 4935 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.080035 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.080108 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aad90193-5bba-4acd-ac25-d67d21a165d3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.080163 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr69b\" (UniqueName: \"kubernetes.io/projected/aad90193-5bba-4acd-ac25-d67d21a165d3-kube-api-access-gr69b\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.270582 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" event={"ID":"aad90193-5bba-4acd-ac25-d67d21a165d3","Type":"ContainerDied","Data":"db452dc4ac9d3592502936c9fa30cf4d5cbd57ad0b345c10364d913e01d6737b"} Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.271301 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db452dc4ac9d3592502936c9fa30cf4d5cbd57ad0b345c10364d913e01d6737b" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.270832 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp" Oct 05 09:08:34 crc kubenswrapper[4935]: I1005 09:08:34.815434 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8" path="/var/lib/kubelet/pods/9ac2b72b-22cb-4d6e-bab9-84a6fde5f2d8/volumes" Oct 05 09:08:44 crc kubenswrapper[4935]: I1005 09:08:44.289514 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:08:44 crc kubenswrapper[4935]: I1005 09:08:44.290140 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:09:02 crc kubenswrapper[4935]: I1005 09:09:02.486524 4935 scope.go:117] "RemoveContainer" containerID="c7e854e4ff1d796d9c50575c54bc8d4800e56ae66443626cb9b95683116536b6" Oct 05 09:09:02 crc kubenswrapper[4935]: I1005 09:09:02.520564 4935 scope.go:117] "RemoveContainer" containerID="6b176d45361a51321576e1b653290714ee40f860bafe31cd3cf256bb538ac401" Oct 05 09:09:02 crc kubenswrapper[4935]: I1005 09:09:02.594114 4935 scope.go:117] "RemoveContainer" containerID="3d59c116227bf71be70d0fd62ba7516148990b16ced050da9a81d91e9e641ab7" Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.289840 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.290451 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.290596 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.291407 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3f5371d7f74bebfc1a9f20169e0444802d8ff98646f3001d261f7f1c9425f2f"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.291466 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://b3f5371d7f74bebfc1a9f20169e0444802d8ff98646f3001d261f7f1c9425f2f" gracePeriod=600 Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.780705 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="b3f5371d7f74bebfc1a9f20169e0444802d8ff98646f3001d261f7f1c9425f2f" exitCode=0 Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.793169 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"b3f5371d7f74bebfc1a9f20169e0444802d8ff98646f3001d261f7f1c9425f2f"} Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.793251 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46"} Oct 05 09:09:14 crc kubenswrapper[4935]: I1005 09:09:14.793281 4935 scope.go:117] "RemoveContainer" containerID="b3f98a137c89eff0ff0c9d273c1c9eb795fdb027762d40b17ab2f9541aa8b652" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.086954 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pqmj6"] Oct 05 09:09:55 crc kubenswrapper[4935]: E1005 09:09:55.088080 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="registry-server" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088098 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="registry-server" Oct 05 09:09:55 crc kubenswrapper[4935]: E1005 09:09:55.088113 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerName="extract-utilities" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088119 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerName="extract-utilities" Oct 05 09:09:55 crc kubenswrapper[4935]: E1005 09:09:55.088137 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="extract-utilities" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088145 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="extract-utilities" Oct 05 09:09:55 crc kubenswrapper[4935]: E1005 09:09:55.088165 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerName="extract-content" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088173 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerName="extract-content" Oct 05 09:09:55 crc kubenswrapper[4935]: E1005 09:09:55.088181 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad90193-5bba-4acd-ac25-d67d21a165d3" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088190 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad90193-5bba-4acd-ac25-d67d21a165d3" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 05 09:09:55 crc kubenswrapper[4935]: E1005 09:09:55.088207 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerName="registry-server" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088216 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerName="registry-server" Oct 05 09:09:55 crc kubenswrapper[4935]: E1005 09:09:55.088239 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="extract-content" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088246 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="extract-content" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088472 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9a4be3-ebdb-459f-a35f-089cfaac3920" containerName="registry-server" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088498 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="78831ce6-216f-46aa-8f3a-dc629b4ccbf1" containerName="registry-server" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.088514 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad90193-5bba-4acd-ac25-d67d21a165d3" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.090371 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.107102 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqmj6"] Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.168382 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f89v8\" (UniqueName: \"kubernetes.io/projected/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-kube-api-access-f89v8\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.168549 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-catalog-content\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.168608 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-utilities\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.272553 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-catalog-content\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.272650 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-utilities\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.272742 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f89v8\" (UniqueName: \"kubernetes.io/projected/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-kube-api-access-f89v8\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.273240 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-catalog-content\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.273351 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-utilities\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.297043 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f89v8\" (UniqueName: \"kubernetes.io/projected/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-kube-api-access-f89v8\") pod \"redhat-marketplace-pqmj6\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.409325 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:09:55 crc kubenswrapper[4935]: I1005 09:09:55.919727 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqmj6"] Oct 05 09:09:56 crc kubenswrapper[4935]: I1005 09:09:56.276507 4935 generic.go:334] "Generic (PLEG): container finished" podID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerID="034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06" exitCode=0 Oct 05 09:09:56 crc kubenswrapper[4935]: I1005 09:09:56.276593 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqmj6" event={"ID":"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862","Type":"ContainerDied","Data":"034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06"} Oct 05 09:09:56 crc kubenswrapper[4935]: I1005 09:09:56.276645 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqmj6" event={"ID":"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862","Type":"ContainerStarted","Data":"78e94a64af2e9e74e455c552159a7dd12219735526fd3bc35cfc5244ea90da50"} Oct 05 09:09:56 crc kubenswrapper[4935]: I1005 09:09:56.279956 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:09:58 crc kubenswrapper[4935]: I1005 09:09:58.318882 4935 generic.go:334] "Generic (PLEG): container finished" podID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerID="c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9" exitCode=0 Oct 05 09:09:58 crc kubenswrapper[4935]: I1005 09:09:58.318991 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqmj6" event={"ID":"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862","Type":"ContainerDied","Data":"c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9"} Oct 05 09:09:59 crc kubenswrapper[4935]: I1005 09:09:59.331709 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqmj6" event={"ID":"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862","Type":"ContainerStarted","Data":"46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15"} Oct 05 09:09:59 crc kubenswrapper[4935]: I1005 09:09:59.354485 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pqmj6" podStartSLOduration=1.871752865 podStartE2EDuration="4.354466108s" podCreationTimestamp="2025-10-05 09:09:55 +0000 UTC" firstStartedPulling="2025-10-05 09:09:56.279350969 +0000 UTC m=+8230.161977469" lastFinishedPulling="2025-10-05 09:09:58.762064212 +0000 UTC m=+8232.644690712" observedRunningTime="2025-10-05 09:09:59.351383826 +0000 UTC m=+8233.234010306" watchObservedRunningTime="2025-10-05 09:09:59.354466108 +0000 UTC m=+8233.237092568" Oct 05 09:10:05 crc kubenswrapper[4935]: I1005 09:10:05.410047 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:10:05 crc kubenswrapper[4935]: I1005 09:10:05.410707 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:10:05 crc kubenswrapper[4935]: I1005 09:10:05.459885 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:10:06 crc kubenswrapper[4935]: I1005 09:10:06.485590 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:10:06 crc kubenswrapper[4935]: I1005 09:10:06.538089 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqmj6"] Oct 05 09:10:08 crc kubenswrapper[4935]: I1005 09:10:08.456569 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pqmj6" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerName="registry-server" containerID="cri-o://46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15" gracePeriod=2 Oct 05 09:10:08 crc kubenswrapper[4935]: I1005 09:10:08.988263 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.115376 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f89v8\" (UniqueName: \"kubernetes.io/projected/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-kube-api-access-f89v8\") pod \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.115444 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-utilities\") pod \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.115599 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-catalog-content\") pod \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\" (UID: \"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862\") " Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.116805 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-utilities" (OuterVolumeSpecName: "utilities") pod "6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" (UID: "6db5af5a-ab7b-4b7f-89b4-e2e419bc4862"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.117548 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.127187 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-kube-api-access-f89v8" (OuterVolumeSpecName: "kube-api-access-f89v8") pod "6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" (UID: "6db5af5a-ab7b-4b7f-89b4-e2e419bc4862"). InnerVolumeSpecName "kube-api-access-f89v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.145186 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" (UID: "6db5af5a-ab7b-4b7f-89b4-e2e419bc4862"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.219882 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f89v8\" (UniqueName: \"kubernetes.io/projected/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-kube-api-access-f89v8\") on node \"crc\" DevicePath \"\"" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.219947 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.475365 4935 generic.go:334] "Generic (PLEG): container finished" podID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerID="46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15" exitCode=0 Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.475425 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqmj6" event={"ID":"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862","Type":"ContainerDied","Data":"46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15"} Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.475489 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqmj6" event={"ID":"6db5af5a-ab7b-4b7f-89b4-e2e419bc4862","Type":"ContainerDied","Data":"78e94a64af2e9e74e455c552159a7dd12219735526fd3bc35cfc5244ea90da50"} Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.475512 4935 scope.go:117] "RemoveContainer" containerID="46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.476101 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqmj6" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.528048 4935 scope.go:117] "RemoveContainer" containerID="c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.534076 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqmj6"] Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.543466 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqmj6"] Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.561194 4935 scope.go:117] "RemoveContainer" containerID="034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.630090 4935 scope.go:117] "RemoveContainer" containerID="46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15" Oct 05 09:10:09 crc kubenswrapper[4935]: E1005 09:10:09.630485 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15\": container with ID starting with 46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15 not found: ID does not exist" containerID="46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.630532 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15"} err="failed to get container status \"46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15\": rpc error: code = NotFound desc = could not find container \"46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15\": container with ID starting with 46ecc9b579ac9f4861f641c654216a9872e194c07b6e643bb725372895a73b15 not found: ID does not exist" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.630561 4935 scope.go:117] "RemoveContainer" containerID="c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9" Oct 05 09:10:09 crc kubenswrapper[4935]: E1005 09:10:09.630871 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9\": container with ID starting with c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9 not found: ID does not exist" containerID="c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.630962 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9"} err="failed to get container status \"c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9\": rpc error: code = NotFound desc = could not find container \"c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9\": container with ID starting with c588680c5e716e3cb744a136a5de0c72a1b8f211a66ec02f970e973ac5f50cf9 not found: ID does not exist" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.630999 4935 scope.go:117] "RemoveContainer" containerID="034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06" Oct 05 09:10:09 crc kubenswrapper[4935]: E1005 09:10:09.631375 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06\": container with ID starting with 034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06 not found: ID does not exist" containerID="034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06" Oct 05 09:10:09 crc kubenswrapper[4935]: I1005 09:10:09.631413 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06"} err="failed to get container status \"034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06\": rpc error: code = NotFound desc = could not find container \"034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06\": container with ID starting with 034dbe5e23f579505d77c7190e59694fc31115e484bc1a355b3e18debfbfbf06 not found: ID does not exist" Oct 05 09:10:10 crc kubenswrapper[4935]: I1005 09:10:10.791398 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" path="/var/lib/kubelet/pods/6db5af5a-ab7b-4b7f-89b4-e2e419bc4862/volumes" Oct 05 09:10:42 crc kubenswrapper[4935]: I1005 09:10:42.058740 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-4bqbw"] Oct 05 09:10:42 crc kubenswrapper[4935]: I1005 09:10:42.077934 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-4bqbw"] Oct 05 09:10:42 crc kubenswrapper[4935]: I1005 09:10:42.791654 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0249f885-bcbd-45f3-bbb6-b1285f4b47d0" path="/var/lib/kubelet/pods/0249f885-bcbd-45f3-bbb6-b1285f4b47d0/volumes" Oct 05 09:10:51 crc kubenswrapper[4935]: I1005 09:10:51.025114 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-35a0-account-create-fhstx"] Oct 05 09:10:51 crc kubenswrapper[4935]: I1005 09:10:51.035064 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-35a0-account-create-fhstx"] Oct 05 09:10:52 crc kubenswrapper[4935]: I1005 09:10:52.792116 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b6dceca-640d-4af2-845d-48741604e927" path="/var/lib/kubelet/pods/0b6dceca-640d-4af2-845d-48741604e927/volumes" Oct 05 09:11:02 crc kubenswrapper[4935]: I1005 09:11:02.763127 4935 scope.go:117] "RemoveContainer" containerID="61ad34ae00274c7a1357ebf1f9f69d5bfb9c5d015b8c87a3a29465393b473ace" Oct 05 09:11:02 crc kubenswrapper[4935]: I1005 09:11:02.797033 4935 scope.go:117] "RemoveContainer" containerID="2128de18a6b393434ef20a05a2786e15dce2a908b0b2f84319d87254057f62ee" Oct 05 09:11:03 crc kubenswrapper[4935]: I1005 09:11:03.033353 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-x2xbd"] Oct 05 09:11:03 crc kubenswrapper[4935]: I1005 09:11:03.044565 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-x2xbd"] Oct 05 09:11:04 crc kubenswrapper[4935]: I1005 09:11:04.789103 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6" path="/var/lib/kubelet/pods/6e0fa5c4-7fee-4b93-8ab1-3eb46da82cc6/volumes" Oct 05 09:11:14 crc kubenswrapper[4935]: I1005 09:11:14.289550 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:11:14 crc kubenswrapper[4935]: I1005 09:11:14.290162 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:11:23 crc kubenswrapper[4935]: I1005 09:11:23.051948 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-kncpn"] Oct 05 09:11:23 crc kubenswrapper[4935]: I1005 09:11:23.059587 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-kncpn"] Oct 05 09:11:24 crc kubenswrapper[4935]: I1005 09:11:24.793201 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72a208bb-52b8-43e3-b8f4-21511674b129" path="/var/lib/kubelet/pods/72a208bb-52b8-43e3-b8f4-21511674b129/volumes" Oct 05 09:11:33 crc kubenswrapper[4935]: I1005 09:11:33.034037 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-00a5-account-create-f7wmx"] Oct 05 09:11:33 crc kubenswrapper[4935]: I1005 09:11:33.044634 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-00a5-account-create-f7wmx"] Oct 05 09:11:34 crc kubenswrapper[4935]: I1005 09:11:34.789882 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6debd557-00e3-424d-b70a-f94f1651f8f0" path="/var/lib/kubelet/pods/6debd557-00e3-424d-b70a-f94f1651f8f0/volumes" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.170218 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wxw4k"] Oct 05 09:11:40 crc kubenswrapper[4935]: E1005 09:11:40.171367 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerName="extract-utilities" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.171387 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerName="extract-utilities" Oct 05 09:11:40 crc kubenswrapper[4935]: E1005 09:11:40.171409 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerName="registry-server" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.171421 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerName="registry-server" Oct 05 09:11:40 crc kubenswrapper[4935]: E1005 09:11:40.171468 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerName="extract-content" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.171479 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerName="extract-content" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.171749 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6db5af5a-ab7b-4b7f-89b4-e2e419bc4862" containerName="registry-server" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.174001 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.175512 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5r7s\" (UniqueName: \"kubernetes.io/projected/df5dd0b6-0401-4998-a489-09399085ec0a-kube-api-access-r5r7s\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.175573 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-utilities\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.175698 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-catalog-content\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.180660 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wxw4k"] Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.278576 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5r7s\" (UniqueName: \"kubernetes.io/projected/df5dd0b6-0401-4998-a489-09399085ec0a-kube-api-access-r5r7s\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.278925 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-utilities\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.279029 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-catalog-content\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.279557 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-utilities\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.279836 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-catalog-content\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.303327 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5r7s\" (UniqueName: \"kubernetes.io/projected/df5dd0b6-0401-4998-a489-09399085ec0a-kube-api-access-r5r7s\") pod \"community-operators-wxw4k\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:40 crc kubenswrapper[4935]: I1005 09:11:40.505800 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:41 crc kubenswrapper[4935]: I1005 09:11:41.028443 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wxw4k"] Oct 05 09:11:41 crc kubenswrapper[4935]: I1005 09:11:41.473901 4935 generic.go:334] "Generic (PLEG): container finished" podID="df5dd0b6-0401-4998-a489-09399085ec0a" containerID="93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a" exitCode=0 Oct 05 09:11:41 crc kubenswrapper[4935]: I1005 09:11:41.474020 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxw4k" event={"ID":"df5dd0b6-0401-4998-a489-09399085ec0a","Type":"ContainerDied","Data":"93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a"} Oct 05 09:11:41 crc kubenswrapper[4935]: I1005 09:11:41.474244 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxw4k" event={"ID":"df5dd0b6-0401-4998-a489-09399085ec0a","Type":"ContainerStarted","Data":"35b6df686c95a253b7787b2072f6141776396bddd24a588ffde2e8b3ae709882"} Oct 05 09:11:43 crc kubenswrapper[4935]: I1005 09:11:43.501303 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxw4k" event={"ID":"df5dd0b6-0401-4998-a489-09399085ec0a","Type":"ContainerStarted","Data":"ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710"} Oct 05 09:11:44 crc kubenswrapper[4935]: I1005 09:11:44.289828 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:11:44 crc kubenswrapper[4935]: I1005 09:11:44.289974 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:11:44 crc kubenswrapper[4935]: I1005 09:11:44.517449 4935 generic.go:334] "Generic (PLEG): container finished" podID="df5dd0b6-0401-4998-a489-09399085ec0a" containerID="ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710" exitCode=0 Oct 05 09:11:44 crc kubenswrapper[4935]: I1005 09:11:44.517566 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxw4k" event={"ID":"df5dd0b6-0401-4998-a489-09399085ec0a","Type":"ContainerDied","Data":"ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710"} Oct 05 09:11:45 crc kubenswrapper[4935]: I1005 09:11:45.060576 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-wn8z7"] Oct 05 09:11:45 crc kubenswrapper[4935]: I1005 09:11:45.073573 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-wn8z7"] Oct 05 09:11:45 crc kubenswrapper[4935]: I1005 09:11:45.532133 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxw4k" event={"ID":"df5dd0b6-0401-4998-a489-09399085ec0a","Type":"ContainerStarted","Data":"bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc"} Oct 05 09:11:45 crc kubenswrapper[4935]: I1005 09:11:45.555248 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wxw4k" podStartSLOduration=2.067482554 podStartE2EDuration="5.555214071s" podCreationTimestamp="2025-10-05 09:11:40 +0000 UTC" firstStartedPulling="2025-10-05 09:11:41.476540593 +0000 UTC m=+8335.359167053" lastFinishedPulling="2025-10-05 09:11:44.96427211 +0000 UTC m=+8338.846898570" observedRunningTime="2025-10-05 09:11:45.551094693 +0000 UTC m=+8339.433721163" watchObservedRunningTime="2025-10-05 09:11:45.555214071 +0000 UTC m=+8339.437840571" Oct 05 09:11:46 crc kubenswrapper[4935]: I1005 09:11:46.792815 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0315ba9d-0e75-4013-b5d0-07728f5d07a0" path="/var/lib/kubelet/pods/0315ba9d-0e75-4013-b5d0-07728f5d07a0/volumes" Oct 05 09:11:50 crc kubenswrapper[4935]: I1005 09:11:50.506246 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:50 crc kubenswrapper[4935]: I1005 09:11:50.506717 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:50 crc kubenswrapper[4935]: I1005 09:11:50.550377 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:50 crc kubenswrapper[4935]: I1005 09:11:50.648767 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:50 crc kubenswrapper[4935]: I1005 09:11:50.801779 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wxw4k"] Oct 05 09:11:52 crc kubenswrapper[4935]: I1005 09:11:52.601111 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wxw4k" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" containerName="registry-server" containerID="cri-o://bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc" gracePeriod=2 Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.136029 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.296063 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-catalog-content\") pod \"df5dd0b6-0401-4998-a489-09399085ec0a\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.296534 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-utilities\") pod \"df5dd0b6-0401-4998-a489-09399085ec0a\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.296650 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5r7s\" (UniqueName: \"kubernetes.io/projected/df5dd0b6-0401-4998-a489-09399085ec0a-kube-api-access-r5r7s\") pod \"df5dd0b6-0401-4998-a489-09399085ec0a\" (UID: \"df5dd0b6-0401-4998-a489-09399085ec0a\") " Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.299339 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-utilities" (OuterVolumeSpecName: "utilities") pod "df5dd0b6-0401-4998-a489-09399085ec0a" (UID: "df5dd0b6-0401-4998-a489-09399085ec0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.306596 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df5dd0b6-0401-4998-a489-09399085ec0a-kube-api-access-r5r7s" (OuterVolumeSpecName: "kube-api-access-r5r7s") pod "df5dd0b6-0401-4998-a489-09399085ec0a" (UID: "df5dd0b6-0401-4998-a489-09399085ec0a"). InnerVolumeSpecName "kube-api-access-r5r7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.399281 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.399312 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5r7s\" (UniqueName: \"kubernetes.io/projected/df5dd0b6-0401-4998-a489-09399085ec0a-kube-api-access-r5r7s\") on node \"crc\" DevicePath \"\"" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.604999 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df5dd0b6-0401-4998-a489-09399085ec0a" (UID: "df5dd0b6-0401-4998-a489-09399085ec0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.605431 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5dd0b6-0401-4998-a489-09399085ec0a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.617317 4935 generic.go:334] "Generic (PLEG): container finished" podID="df5dd0b6-0401-4998-a489-09399085ec0a" containerID="bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc" exitCode=0 Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.617385 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxw4k" event={"ID":"df5dd0b6-0401-4998-a489-09399085ec0a","Type":"ContainerDied","Data":"bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc"} Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.617441 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxw4k" event={"ID":"df5dd0b6-0401-4998-a489-09399085ec0a","Type":"ContainerDied","Data":"35b6df686c95a253b7787b2072f6141776396bddd24a588ffde2e8b3ae709882"} Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.617471 4935 scope.go:117] "RemoveContainer" containerID="bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.617468 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wxw4k" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.647315 4935 scope.go:117] "RemoveContainer" containerID="ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.673129 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wxw4k"] Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.689063 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wxw4k"] Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.719039 4935 scope.go:117] "RemoveContainer" containerID="93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.751458 4935 scope.go:117] "RemoveContainer" containerID="bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc" Oct 05 09:11:53 crc kubenswrapper[4935]: E1005 09:11:53.751950 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc\": container with ID starting with bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc not found: ID does not exist" containerID="bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.751989 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc"} err="failed to get container status \"bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc\": rpc error: code = NotFound desc = could not find container \"bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc\": container with ID starting with bef051d29ad5fa8918f1628c479dbfafc433a10879470a92d4876881f4c77cbc not found: ID does not exist" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.752017 4935 scope.go:117] "RemoveContainer" containerID="ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710" Oct 05 09:11:53 crc kubenswrapper[4935]: E1005 09:11:53.752326 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710\": container with ID starting with ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710 not found: ID does not exist" containerID="ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.752352 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710"} err="failed to get container status \"ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710\": rpc error: code = NotFound desc = could not find container \"ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710\": container with ID starting with ca0c86dbf9706018fe8907ebf8fd09c7f26e1dbb18126fcb7d35b7967e89c710 not found: ID does not exist" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.752370 4935 scope.go:117] "RemoveContainer" containerID="93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a" Oct 05 09:11:53 crc kubenswrapper[4935]: E1005 09:11:53.753189 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a\": container with ID starting with 93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a not found: ID does not exist" containerID="93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a" Oct 05 09:11:53 crc kubenswrapper[4935]: I1005 09:11:53.753224 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a"} err="failed to get container status \"93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a\": rpc error: code = NotFound desc = could not find container \"93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a\": container with ID starting with 93bddc7f5869325ab423ecc23777836b8fa815d2cbd7a4f1c0509478cee7568a not found: ID does not exist" Oct 05 09:11:54 crc kubenswrapper[4935]: I1005 09:11:54.797701 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" path="/var/lib/kubelet/pods/df5dd0b6-0401-4998-a489-09399085ec0a/volumes" Oct 05 09:11:55 crc kubenswrapper[4935]: I1005 09:11:55.644700 4935 generic.go:334] "Generic (PLEG): container finished" podID="097a7b90-7e36-4aa4-b9df-e17362922f55" containerID="2035fcb4cf084fe243d6116a7c4b54048adb85ac94aad7a62adcaa6a9eb0f0a9" exitCode=0 Oct 05 09:11:55 crc kubenswrapper[4935]: I1005 09:11:55.644749 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" event={"ID":"097a7b90-7e36-4aa4-b9df-e17362922f55","Type":"ContainerDied","Data":"2035fcb4cf084fe243d6116a7c4b54048adb85ac94aad7a62adcaa6a9eb0f0a9"} Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.115285 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.195746 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-inventory\") pod \"097a7b90-7e36-4aa4-b9df-e17362922f55\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.195934 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ceph\") pod \"097a7b90-7e36-4aa4-b9df-e17362922f55\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.195974 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np8gt\" (UniqueName: \"kubernetes.io/projected/097a7b90-7e36-4aa4-b9df-e17362922f55-kube-api-access-np8gt\") pod \"097a7b90-7e36-4aa4-b9df-e17362922f55\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.196025 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ssh-key\") pod \"097a7b90-7e36-4aa4-b9df-e17362922f55\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.196067 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-tripleo-cleanup-combined-ca-bundle\") pod \"097a7b90-7e36-4aa4-b9df-e17362922f55\" (UID: \"097a7b90-7e36-4aa4-b9df-e17362922f55\") " Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.202041 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ceph" (OuterVolumeSpecName: "ceph") pod "097a7b90-7e36-4aa4-b9df-e17362922f55" (UID: "097a7b90-7e36-4aa4-b9df-e17362922f55"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.202062 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "097a7b90-7e36-4aa4-b9df-e17362922f55" (UID: "097a7b90-7e36-4aa4-b9df-e17362922f55"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.202078 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097a7b90-7e36-4aa4-b9df-e17362922f55-kube-api-access-np8gt" (OuterVolumeSpecName: "kube-api-access-np8gt") pod "097a7b90-7e36-4aa4-b9df-e17362922f55" (UID: "097a7b90-7e36-4aa4-b9df-e17362922f55"). InnerVolumeSpecName "kube-api-access-np8gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.224060 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "097a7b90-7e36-4aa4-b9df-e17362922f55" (UID: "097a7b90-7e36-4aa4-b9df-e17362922f55"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.229930 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-inventory" (OuterVolumeSpecName: "inventory") pod "097a7b90-7e36-4aa4-b9df-e17362922f55" (UID: "097a7b90-7e36-4aa4-b9df-e17362922f55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.298455 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.298490 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.298505 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np8gt\" (UniqueName: \"kubernetes.io/projected/097a7b90-7e36-4aa4-b9df-e17362922f55-kube-api-access-np8gt\") on node \"crc\" DevicePath \"\"" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.298522 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.298538 4935 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097a7b90-7e36-4aa4-b9df-e17362922f55-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.677079 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" event={"ID":"097a7b90-7e36-4aa4-b9df-e17362922f55","Type":"ContainerDied","Data":"707e621d1347ca891aed52e46b35671344ca7c6c503380cf0f4c188a56322a09"} Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.677152 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="707e621d1347ca891aed52e46b35671344ca7c6c503380cf0f4c188a56322a09" Oct 05 09:11:57 crc kubenswrapper[4935]: I1005 09:11:57.677258 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx" Oct 05 09:12:02 crc kubenswrapper[4935]: I1005 09:12:02.899916 4935 scope.go:117] "RemoveContainer" containerID="220670d55e5cfad5d7b830c25e329103c382da0fb77952d4cec10e7549a4b9ea" Oct 05 09:12:02 crc kubenswrapper[4935]: I1005 09:12:02.939791 4935 scope.go:117] "RemoveContainer" containerID="638c462666eed8bd0c52130cadcebd82727a70fddb0eb0ebf09d0d6918c5cd3f" Oct 05 09:12:02 crc kubenswrapper[4935]: I1005 09:12:02.977841 4935 scope.go:117] "RemoveContainer" containerID="3f6ff4ea375bafaa059943b0faae2faee9d9728eb79375bda1e812f961bce74b" Oct 05 09:12:03 crc kubenswrapper[4935]: I1005 09:12:03.030546 4935 scope.go:117] "RemoveContainer" containerID="d60eeb5c910f425f5c1afe5c3640ca78036735a2a16d87ec97b6cb4cf77394b2" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.326800 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rtkmw"] Oct 05 09:12:06 crc kubenswrapper[4935]: E1005 09:12:06.327685 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" containerName="extract-utilities" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.327698 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" containerName="extract-utilities" Oct 05 09:12:06 crc kubenswrapper[4935]: E1005 09:12:06.327718 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" containerName="extract-content" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.327725 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" containerName="extract-content" Oct 05 09:12:06 crc kubenswrapper[4935]: E1005 09:12:06.327744 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" containerName="registry-server" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.327750 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" containerName="registry-server" Oct 05 09:12:06 crc kubenswrapper[4935]: E1005 09:12:06.327758 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097a7b90-7e36-4aa4-b9df-e17362922f55" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.327766 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="097a7b90-7e36-4aa4-b9df-e17362922f55" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.327985 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="097a7b90-7e36-4aa4-b9df-e17362922f55" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.327997 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="df5dd0b6-0401-4998-a489-09399085ec0a" containerName="registry-server" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.328708 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.336336 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.336801 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.337049 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.337210 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.354716 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-bk27c"] Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.356142 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.357876 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.358072 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.367383 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rtkmw"] Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.380445 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-bk27c"] Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519207 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519282 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519307 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-inventory\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519413 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6lnr\" (UniqueName: \"kubernetes.io/projected/6870738c-a0fd-4448-b129-be1246e47542-kube-api-access-x6lnr\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519450 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-inventory\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519509 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519738 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-ssh-key\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519848 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ceph\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.519946 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjhj6\" (UniqueName: \"kubernetes.io/projected/3da548b8-5970-4e0c-9b11-261950cb6e8d-kube-api-access-hjhj6\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.621534 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.621593 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-ssh-key\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.621614 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ceph\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.621647 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjhj6\" (UniqueName: \"kubernetes.io/projected/3da548b8-5970-4e0c-9b11-261950cb6e8d-kube-api-access-hjhj6\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.621706 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.621737 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.622483 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-inventory\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.622548 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6lnr\" (UniqueName: \"kubernetes.io/projected/6870738c-a0fd-4448-b129-be1246e47542-kube-api-access-x6lnr\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.622579 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-inventory\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.627401 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-inventory\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.627688 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ceph\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.628230 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.628639 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.628998 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-inventory\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.631415 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-ssh-key\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.636909 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.639165 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6lnr\" (UniqueName: \"kubernetes.io/projected/6870738c-a0fd-4448-b129-be1246e47542-kube-api-access-x6lnr\") pod \"bootstrap-openstack-openstack-networker-bk27c\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.641976 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjhj6\" (UniqueName: \"kubernetes.io/projected/3da548b8-5970-4e0c-9b11-261950cb6e8d-kube-api-access-hjhj6\") pod \"bootstrap-openstack-openstack-cell1-rtkmw\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.647151 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:12:06 crc kubenswrapper[4935]: I1005 09:12:06.680106 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:12:07 crc kubenswrapper[4935]: I1005 09:12:07.204477 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rtkmw"] Oct 05 09:12:07 crc kubenswrapper[4935]: I1005 09:12:07.320490 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-bk27c"] Oct 05 09:12:07 crc kubenswrapper[4935]: W1005 09:12:07.328469 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6870738c_a0fd_4448_b129_be1246e47542.slice/crio-e874f3d0735298276e734d2cff638d99edcdf9298a67355470522510094367a0 WatchSource:0}: Error finding container e874f3d0735298276e734d2cff638d99edcdf9298a67355470522510094367a0: Status 404 returned error can't find the container with id e874f3d0735298276e734d2cff638d99edcdf9298a67355470522510094367a0 Oct 05 09:12:07 crc kubenswrapper[4935]: I1005 09:12:07.841615 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" event={"ID":"6870738c-a0fd-4448-b129-be1246e47542","Type":"ContainerStarted","Data":"e874f3d0735298276e734d2cff638d99edcdf9298a67355470522510094367a0"} Oct 05 09:12:07 crc kubenswrapper[4935]: I1005 09:12:07.843843 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" event={"ID":"3da548b8-5970-4e0c-9b11-261950cb6e8d","Type":"ContainerStarted","Data":"2238b9d76902794de11d091ac9a65050efbb8312d45ff72b0207bde787f93b77"} Oct 05 09:12:08 crc kubenswrapper[4935]: I1005 09:12:08.855785 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" event={"ID":"3da548b8-5970-4e0c-9b11-261950cb6e8d","Type":"ContainerStarted","Data":"8b2302423d870ff900ac34d126e9772e7a1409fb52f5981326bfd40664c0d5e0"} Oct 05 09:12:08 crc kubenswrapper[4935]: I1005 09:12:08.858371 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" event={"ID":"6870738c-a0fd-4448-b129-be1246e47542","Type":"ContainerStarted","Data":"21435b343948351841b4761a458d74e0f566130e5e08eb676d8c9866b46dc682"} Oct 05 09:12:08 crc kubenswrapper[4935]: I1005 09:12:08.878698 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" podStartSLOduration=2.377570843 podStartE2EDuration="2.878663492s" podCreationTimestamp="2025-10-05 09:12:06 +0000 UTC" firstStartedPulling="2025-10-05 09:12:07.202331734 +0000 UTC m=+8361.084958194" lastFinishedPulling="2025-10-05 09:12:07.703424383 +0000 UTC m=+8361.586050843" observedRunningTime="2025-10-05 09:12:08.869987104 +0000 UTC m=+8362.752613574" watchObservedRunningTime="2025-10-05 09:12:08.878663492 +0000 UTC m=+8362.761289962" Oct 05 09:12:08 crc kubenswrapper[4935]: I1005 09:12:08.895055 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" podStartSLOduration=2.429917848 podStartE2EDuration="2.895027602s" podCreationTimestamp="2025-10-05 09:12:06 +0000 UTC" firstStartedPulling="2025-10-05 09:12:07.330768389 +0000 UTC m=+8361.213394859" lastFinishedPulling="2025-10-05 09:12:07.795878113 +0000 UTC m=+8361.678504613" observedRunningTime="2025-10-05 09:12:08.886642901 +0000 UTC m=+8362.769269381" watchObservedRunningTime="2025-10-05 09:12:08.895027602 +0000 UTC m=+8362.777654082" Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.289671 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.290331 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.290420 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.291596 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.291724 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" gracePeriod=600 Oct 05 09:12:14 crc kubenswrapper[4935]: E1005 09:12:14.423046 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.930563 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" exitCode=0 Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.930648 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46"} Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.931082 4935 scope.go:117] "RemoveContainer" containerID="b3f5371d7f74bebfc1a9f20169e0444802d8ff98646f3001d261f7f1c9425f2f" Oct 05 09:12:14 crc kubenswrapper[4935]: I1005 09:12:14.932497 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:12:14 crc kubenswrapper[4935]: E1005 09:12:14.932802 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:12:26 crc kubenswrapper[4935]: I1005 09:12:26.785949 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:12:26 crc kubenswrapper[4935]: E1005 09:12:26.786846 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:12:39 crc kubenswrapper[4935]: I1005 09:12:39.776876 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:12:39 crc kubenswrapper[4935]: E1005 09:12:39.777685 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:12:52 crc kubenswrapper[4935]: I1005 09:12:52.777034 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:12:52 crc kubenswrapper[4935]: E1005 09:12:52.777985 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:13:03 crc kubenswrapper[4935]: I1005 09:13:03.777479 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:13:03 crc kubenswrapper[4935]: E1005 09:13:03.778923 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:13:14 crc kubenswrapper[4935]: I1005 09:13:14.777777 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:13:14 crc kubenswrapper[4935]: E1005 09:13:14.778708 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:13:28 crc kubenswrapper[4935]: I1005 09:13:28.779029 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:13:28 crc kubenswrapper[4935]: E1005 09:13:28.780435 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:13:41 crc kubenswrapper[4935]: I1005 09:13:41.777298 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:13:41 crc kubenswrapper[4935]: E1005 09:13:41.778326 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:13:56 crc kubenswrapper[4935]: I1005 09:13:56.789591 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:13:56 crc kubenswrapper[4935]: E1005 09:13:56.790877 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:14:08 crc kubenswrapper[4935]: I1005 09:14:08.778299 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:14:08 crc kubenswrapper[4935]: E1005 09:14:08.779524 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:14:21 crc kubenswrapper[4935]: I1005 09:14:21.776777 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:14:21 crc kubenswrapper[4935]: E1005 09:14:21.777662 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:14:35 crc kubenswrapper[4935]: I1005 09:14:35.778735 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:14:35 crc kubenswrapper[4935]: E1005 09:14:35.780248 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:14:50 crc kubenswrapper[4935]: I1005 09:14:50.777193 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:14:50 crc kubenswrapper[4935]: E1005 09:14:50.778014 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:14:54 crc kubenswrapper[4935]: I1005 09:14:54.724413 4935 generic.go:334] "Generic (PLEG): container finished" podID="3da548b8-5970-4e0c-9b11-261950cb6e8d" containerID="8b2302423d870ff900ac34d126e9772e7a1409fb52f5981326bfd40664c0d5e0" exitCode=0 Oct 05 09:14:54 crc kubenswrapper[4935]: I1005 09:14:54.724538 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" event={"ID":"3da548b8-5970-4e0c-9b11-261950cb6e8d","Type":"ContainerDied","Data":"8b2302423d870ff900ac34d126e9772e7a1409fb52f5981326bfd40664c0d5e0"} Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.271339 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.401177 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjhj6\" (UniqueName: \"kubernetes.io/projected/3da548b8-5970-4e0c-9b11-261950cb6e8d-kube-api-access-hjhj6\") pod \"3da548b8-5970-4e0c-9b11-261950cb6e8d\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.401692 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ceph\") pod \"3da548b8-5970-4e0c-9b11-261950cb6e8d\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.401877 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-inventory\") pod \"3da548b8-5970-4e0c-9b11-261950cb6e8d\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.402045 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ssh-key\") pod \"3da548b8-5970-4e0c-9b11-261950cb6e8d\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.402174 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-bootstrap-combined-ca-bundle\") pod \"3da548b8-5970-4e0c-9b11-261950cb6e8d\" (UID: \"3da548b8-5970-4e0c-9b11-261950cb6e8d\") " Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.408468 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3da548b8-5970-4e0c-9b11-261950cb6e8d" (UID: "3da548b8-5970-4e0c-9b11-261950cb6e8d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.408519 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3da548b8-5970-4e0c-9b11-261950cb6e8d-kube-api-access-hjhj6" (OuterVolumeSpecName: "kube-api-access-hjhj6") pod "3da548b8-5970-4e0c-9b11-261950cb6e8d" (UID: "3da548b8-5970-4e0c-9b11-261950cb6e8d"). InnerVolumeSpecName "kube-api-access-hjhj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.413075 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ceph" (OuterVolumeSpecName: "ceph") pod "3da548b8-5970-4e0c-9b11-261950cb6e8d" (UID: "3da548b8-5970-4e0c-9b11-261950cb6e8d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.433778 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3da548b8-5970-4e0c-9b11-261950cb6e8d" (UID: "3da548b8-5970-4e0c-9b11-261950cb6e8d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.450396 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-inventory" (OuterVolumeSpecName: "inventory") pod "3da548b8-5970-4e0c-9b11-261950cb6e8d" (UID: "3da548b8-5970-4e0c-9b11-261950cb6e8d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.505060 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjhj6\" (UniqueName: \"kubernetes.io/projected/3da548b8-5970-4e0c-9b11-261950cb6e8d-kube-api-access-hjhj6\") on node \"crc\" DevicePath \"\"" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.505094 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.505105 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.505115 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.505125 4935 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3da548b8-5970-4e0c-9b11-261950cb6e8d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.751108 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" event={"ID":"3da548b8-5970-4e0c-9b11-261950cb6e8d","Type":"ContainerDied","Data":"2238b9d76902794de11d091ac9a65050efbb8312d45ff72b0207bde787f93b77"} Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.751147 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2238b9d76902794de11d091ac9a65050efbb8312d45ff72b0207bde787f93b77" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.751201 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rtkmw" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.861824 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-ssp86"] Oct 05 09:14:56 crc kubenswrapper[4935]: E1005 09:14:56.862254 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da548b8-5970-4e0c-9b11-261950cb6e8d" containerName="bootstrap-openstack-openstack-cell1" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.862270 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da548b8-5970-4e0c-9b11-261950cb6e8d" containerName="bootstrap-openstack-openstack-cell1" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.862468 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="3da548b8-5970-4e0c-9b11-261950cb6e8d" containerName="bootstrap-openstack-openstack-cell1" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.863225 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.865509 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.874403 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-ssp86"] Oct 05 09:14:56 crc kubenswrapper[4935]: I1005 09:14:56.896396 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.014642 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk2r2\" (UniqueName: \"kubernetes.io/projected/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-kube-api-access-gk2r2\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.014766 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ceph\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.014791 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ssh-key\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.014914 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-inventory\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.116905 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk2r2\" (UniqueName: \"kubernetes.io/projected/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-kube-api-access-gk2r2\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.117054 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ceph\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.117083 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ssh-key\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.117130 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-inventory\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.122559 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ssh-key\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.123005 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ceph\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.133995 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk2r2\" (UniqueName: \"kubernetes.io/projected/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-kube-api-access-gk2r2\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.134409 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-inventory\") pod \"download-cache-openstack-openstack-cell1-ssp86\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.220542 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.797803 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-ssp86"] Oct 05 09:14:57 crc kubenswrapper[4935]: I1005 09:14:57.802845 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:14:58 crc kubenswrapper[4935]: I1005 09:14:58.773757 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" event={"ID":"e39fcf7d-8ea6-4a0a-87ed-a6000793d030","Type":"ContainerStarted","Data":"d484ae8f39fe144b8817f23815ae50472d12f079a7102606bbb3138d72cf33f8"} Oct 05 09:14:58 crc kubenswrapper[4935]: I1005 09:14:58.774223 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" event={"ID":"e39fcf7d-8ea6-4a0a-87ed-a6000793d030","Type":"ContainerStarted","Data":"78f79666665b36ff35fcd9945a8dd179aa1223fc898f7efcf8517bf6576964d7"} Oct 05 09:14:58 crc kubenswrapper[4935]: I1005 09:14:58.794322 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" podStartSLOduration=2.162091286 podStartE2EDuration="2.794298932s" podCreationTimestamp="2025-10-05 09:14:56 +0000 UTC" firstStartedPulling="2025-10-05 09:14:57.802660899 +0000 UTC m=+8531.685287349" lastFinishedPulling="2025-10-05 09:14:58.434868525 +0000 UTC m=+8532.317494995" observedRunningTime="2025-10-05 09:14:58.7926918 +0000 UTC m=+8532.675318260" watchObservedRunningTime="2025-10-05 09:14:58.794298932 +0000 UTC m=+8532.676925402" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.144908 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4"] Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.148495 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.150678 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.151286 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.166825 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4"] Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.296612 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4w42\" (UniqueName: \"kubernetes.io/projected/e72564f1-aaaf-4cdc-8221-2842954a46b7-kube-api-access-f4w42\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.296756 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e72564f1-aaaf-4cdc-8221-2842954a46b7-secret-volume\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.296801 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e72564f1-aaaf-4cdc-8221-2842954a46b7-config-volume\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.398383 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e72564f1-aaaf-4cdc-8221-2842954a46b7-config-volume\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.398528 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4w42\" (UniqueName: \"kubernetes.io/projected/e72564f1-aaaf-4cdc-8221-2842954a46b7-kube-api-access-f4w42\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.398650 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e72564f1-aaaf-4cdc-8221-2842954a46b7-secret-volume\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.399281 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e72564f1-aaaf-4cdc-8221-2842954a46b7-config-volume\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.417026 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e72564f1-aaaf-4cdc-8221-2842954a46b7-secret-volume\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.419361 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4w42\" (UniqueName: \"kubernetes.io/projected/e72564f1-aaaf-4cdc-8221-2842954a46b7-kube-api-access-f4w42\") pod \"collect-profiles-29327595-rfhm4\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.475061 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.946507 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4"] Oct 05 09:15:00 crc kubenswrapper[4935]: I1005 09:15:00.997512 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p2d2q"] Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.000614 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.007777 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p2d2q"] Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.118011 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-catalog-content\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.118463 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-utilities\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.118537 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p67ms\" (UniqueName: \"kubernetes.io/projected/be733bf3-217f-4f6d-8a22-5ab12b4c2528-kube-api-access-p67ms\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.220262 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-catalog-content\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.220448 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-utilities\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.220481 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p67ms\" (UniqueName: \"kubernetes.io/projected/be733bf3-217f-4f6d-8a22-5ab12b4c2528-kube-api-access-p67ms\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.220953 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-catalog-content\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.221353 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-utilities\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.244277 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p67ms\" (UniqueName: \"kubernetes.io/projected/be733bf3-217f-4f6d-8a22-5ab12b4c2528-kube-api-access-p67ms\") pod \"redhat-operators-p2d2q\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.337543 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.846430 4935 generic.go:334] "Generic (PLEG): container finished" podID="e72564f1-aaaf-4cdc-8221-2842954a46b7" containerID="94b5e9de84b890ea9584cd274e0065fc43d37db076c0d92aaaf87bdf7a92eb5c" exitCode=0 Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.846542 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" event={"ID":"e72564f1-aaaf-4cdc-8221-2842954a46b7","Type":"ContainerDied","Data":"94b5e9de84b890ea9584cd274e0065fc43d37db076c0d92aaaf87bdf7a92eb5c"} Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.846715 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" event={"ID":"e72564f1-aaaf-4cdc-8221-2842954a46b7","Type":"ContainerStarted","Data":"d28664ca92845929b0b971fc7926676eb606b0c38cc86e47e9ff6e72d4ae1fa6"} Oct 05 09:15:01 crc kubenswrapper[4935]: I1005 09:15:01.886576 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p2d2q"] Oct 05 09:15:01 crc kubenswrapper[4935]: W1005 09:15:01.890757 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe733bf3_217f_4f6d_8a22_5ab12b4c2528.slice/crio-c1ad823677aead39424c6a1459a847b4b609d450385562bac0538071e364e0c8 WatchSource:0}: Error finding container c1ad823677aead39424c6a1459a847b4b609d450385562bac0538071e364e0c8: Status 404 returned error can't find the container with id c1ad823677aead39424c6a1459a847b4b609d450385562bac0538071e364e0c8 Oct 05 09:15:02 crc kubenswrapper[4935]: I1005 09:15:02.777627 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:15:02 crc kubenswrapper[4935]: E1005 09:15:02.778268 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:15:02 crc kubenswrapper[4935]: I1005 09:15:02.855615 4935 generic.go:334] "Generic (PLEG): container finished" podID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerID="119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b" exitCode=0 Oct 05 09:15:02 crc kubenswrapper[4935]: I1005 09:15:02.855688 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2d2q" event={"ID":"be733bf3-217f-4f6d-8a22-5ab12b4c2528","Type":"ContainerDied","Data":"119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b"} Oct 05 09:15:02 crc kubenswrapper[4935]: I1005 09:15:02.855714 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2d2q" event={"ID":"be733bf3-217f-4f6d-8a22-5ab12b4c2528","Type":"ContainerStarted","Data":"c1ad823677aead39424c6a1459a847b4b609d450385562bac0538071e364e0c8"} Oct 05 09:15:02 crc kubenswrapper[4935]: I1005 09:15:02.857720 4935 generic.go:334] "Generic (PLEG): container finished" podID="6870738c-a0fd-4448-b129-be1246e47542" containerID="21435b343948351841b4761a458d74e0f566130e5e08eb676d8c9866b46dc682" exitCode=0 Oct 05 09:15:02 crc kubenswrapper[4935]: I1005 09:15:02.857931 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" event={"ID":"6870738c-a0fd-4448-b129-be1246e47542","Type":"ContainerDied","Data":"21435b343948351841b4761a458d74e0f566130e5e08eb676d8c9866b46dc682"} Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.272798 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.373112 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e72564f1-aaaf-4cdc-8221-2842954a46b7-secret-volume\") pod \"e72564f1-aaaf-4cdc-8221-2842954a46b7\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.373185 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e72564f1-aaaf-4cdc-8221-2842954a46b7-config-volume\") pod \"e72564f1-aaaf-4cdc-8221-2842954a46b7\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.373265 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4w42\" (UniqueName: \"kubernetes.io/projected/e72564f1-aaaf-4cdc-8221-2842954a46b7-kube-api-access-f4w42\") pod \"e72564f1-aaaf-4cdc-8221-2842954a46b7\" (UID: \"e72564f1-aaaf-4cdc-8221-2842954a46b7\") " Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.375080 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e72564f1-aaaf-4cdc-8221-2842954a46b7-config-volume" (OuterVolumeSpecName: "config-volume") pod "e72564f1-aaaf-4cdc-8221-2842954a46b7" (UID: "e72564f1-aaaf-4cdc-8221-2842954a46b7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.379110 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e72564f1-aaaf-4cdc-8221-2842954a46b7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e72564f1-aaaf-4cdc-8221-2842954a46b7" (UID: "e72564f1-aaaf-4cdc-8221-2842954a46b7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.388168 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e72564f1-aaaf-4cdc-8221-2842954a46b7-kube-api-access-f4w42" (OuterVolumeSpecName: "kube-api-access-f4w42") pod "e72564f1-aaaf-4cdc-8221-2842954a46b7" (UID: "e72564f1-aaaf-4cdc-8221-2842954a46b7"). InnerVolumeSpecName "kube-api-access-f4w42". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.475635 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e72564f1-aaaf-4cdc-8221-2842954a46b7-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.475676 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e72564f1-aaaf-4cdc-8221-2842954a46b7-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.475690 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4w42\" (UniqueName: \"kubernetes.io/projected/e72564f1-aaaf-4cdc-8221-2842954a46b7-kube-api-access-f4w42\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.868221 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" event={"ID":"e72564f1-aaaf-4cdc-8221-2842954a46b7","Type":"ContainerDied","Data":"d28664ca92845929b0b971fc7926676eb606b0c38cc86e47e9ff6e72d4ae1fa6"} Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.868263 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4" Oct 05 09:15:03 crc kubenswrapper[4935]: I1005 09:15:03.868285 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d28664ca92845929b0b971fc7926676eb606b0c38cc86e47e9ff6e72d4ae1fa6" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.355421 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95"] Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.365360 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-zcp95"] Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.420010 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.502062 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-inventory\") pod \"6870738c-a0fd-4448-b129-be1246e47542\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.502169 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6lnr\" (UniqueName: \"kubernetes.io/projected/6870738c-a0fd-4448-b129-be1246e47542-kube-api-access-x6lnr\") pod \"6870738c-a0fd-4448-b129-be1246e47542\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.502223 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-bootstrap-combined-ca-bundle\") pod \"6870738c-a0fd-4448-b129-be1246e47542\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.502414 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-ssh-key\") pod \"6870738c-a0fd-4448-b129-be1246e47542\" (UID: \"6870738c-a0fd-4448-b129-be1246e47542\") " Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.507667 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6870738c-a0fd-4448-b129-be1246e47542" (UID: "6870738c-a0fd-4448-b129-be1246e47542"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.507918 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6870738c-a0fd-4448-b129-be1246e47542-kube-api-access-x6lnr" (OuterVolumeSpecName: "kube-api-access-x6lnr") pod "6870738c-a0fd-4448-b129-be1246e47542" (UID: "6870738c-a0fd-4448-b129-be1246e47542"). InnerVolumeSpecName "kube-api-access-x6lnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.528251 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6870738c-a0fd-4448-b129-be1246e47542" (UID: "6870738c-a0fd-4448-b129-be1246e47542"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.556429 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-inventory" (OuterVolumeSpecName: "inventory") pod "6870738c-a0fd-4448-b129-be1246e47542" (UID: "6870738c-a0fd-4448-b129-be1246e47542"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.606410 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.606713 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6lnr\" (UniqueName: \"kubernetes.io/projected/6870738c-a0fd-4448-b129-be1246e47542-kube-api-access-x6lnr\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.606731 4935 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.606745 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6870738c-a0fd-4448-b129-be1246e47542-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.790647 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99619e61-ede2-432a-a293-5fab27d7d60c" path="/var/lib/kubelet/pods/99619e61-ede2-432a-a293-5fab27d7d60c/volumes" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.879237 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" event={"ID":"6870738c-a0fd-4448-b129-be1246e47542","Type":"ContainerDied","Data":"e874f3d0735298276e734d2cff638d99edcdf9298a67355470522510094367a0"} Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.879328 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e874f3d0735298276e734d2cff638d99edcdf9298a67355470522510094367a0" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.879263 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-bk27c" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.881368 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2d2q" event={"ID":"be733bf3-217f-4f6d-8a22-5ab12b4c2528","Type":"ContainerStarted","Data":"eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62"} Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.964597 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-networker-blbmn"] Oct 05 09:15:04 crc kubenswrapper[4935]: E1005 09:15:04.965049 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e72564f1-aaaf-4cdc-8221-2842954a46b7" containerName="collect-profiles" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.965067 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e72564f1-aaaf-4cdc-8221-2842954a46b7" containerName="collect-profiles" Oct 05 09:15:04 crc kubenswrapper[4935]: E1005 09:15:04.965113 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6870738c-a0fd-4448-b129-be1246e47542" containerName="bootstrap-openstack-openstack-networker" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.965121 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6870738c-a0fd-4448-b129-be1246e47542" containerName="bootstrap-openstack-openstack-networker" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.965305 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e72564f1-aaaf-4cdc-8221-2842954a46b7" containerName="collect-profiles" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.965327 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6870738c-a0fd-4448-b129-be1246e47542" containerName="bootstrap-openstack-openstack-networker" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.966077 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.973792 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.974120 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:15:04 crc kubenswrapper[4935]: I1005 09:15:04.974436 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-blbmn"] Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.116156 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-inventory\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.116571 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-ssh-key\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.116697 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skjbw\" (UniqueName: \"kubernetes.io/projected/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-kube-api-access-skjbw\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.219725 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-ssh-key\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.219810 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skjbw\" (UniqueName: \"kubernetes.io/projected/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-kube-api-access-skjbw\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.219875 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-inventory\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.224028 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-inventory\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.224316 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-ssh-key\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.246331 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skjbw\" (UniqueName: \"kubernetes.io/projected/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-kube-api-access-skjbw\") pod \"download-cache-openstack-openstack-networker-blbmn\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.285391 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.834863 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-blbmn"] Oct 05 09:15:05 crc kubenswrapper[4935]: W1005 09:15:05.848764 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d1fcc45_8bfe_4406_ab6d_fa30a85985cc.slice/crio-4a78e411fd7dac0ca03d8f2b3ca6a0a121f09b27a683d7e264af3f2d20628acb WatchSource:0}: Error finding container 4a78e411fd7dac0ca03d8f2b3ca6a0a121f09b27a683d7e264af3f2d20628acb: Status 404 returned error can't find the container with id 4a78e411fd7dac0ca03d8f2b3ca6a0a121f09b27a683d7e264af3f2d20628acb Oct 05 09:15:05 crc kubenswrapper[4935]: I1005 09:15:05.892912 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-blbmn" event={"ID":"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc","Type":"ContainerStarted","Data":"4a78e411fd7dac0ca03d8f2b3ca6a0a121f09b27a683d7e264af3f2d20628acb"} Oct 05 09:15:06 crc kubenswrapper[4935]: I1005 09:15:06.902342 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-blbmn" event={"ID":"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc","Type":"ContainerStarted","Data":"8ef49d37538c2383d8c833fdf21a0736f23e632980d0ee0721670b862cc98628"} Oct 05 09:15:06 crc kubenswrapper[4935]: I1005 09:15:06.920226 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-networker-blbmn" podStartSLOduration=2.24170989 podStartE2EDuration="2.920174711s" podCreationTimestamp="2025-10-05 09:15:04 +0000 UTC" firstStartedPulling="2025-10-05 09:15:05.851437052 +0000 UTC m=+8539.734063512" lastFinishedPulling="2025-10-05 09:15:06.529901873 +0000 UTC m=+8540.412528333" observedRunningTime="2025-10-05 09:15:06.917291975 +0000 UTC m=+8540.799918435" watchObservedRunningTime="2025-10-05 09:15:06.920174711 +0000 UTC m=+8540.802801181" Oct 05 09:15:07 crc kubenswrapper[4935]: I1005 09:15:07.917258 4935 generic.go:334] "Generic (PLEG): container finished" podID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerID="eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62" exitCode=0 Oct 05 09:15:07 crc kubenswrapper[4935]: I1005 09:15:07.917359 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2d2q" event={"ID":"be733bf3-217f-4f6d-8a22-5ab12b4c2528","Type":"ContainerDied","Data":"eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62"} Oct 05 09:15:08 crc kubenswrapper[4935]: I1005 09:15:08.931719 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2d2q" event={"ID":"be733bf3-217f-4f6d-8a22-5ab12b4c2528","Type":"ContainerStarted","Data":"f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524"} Oct 05 09:15:08 crc kubenswrapper[4935]: I1005 09:15:08.958790 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p2d2q" podStartSLOduration=3.274136674 podStartE2EDuration="8.958773951s" podCreationTimestamp="2025-10-05 09:15:00 +0000 UTC" firstStartedPulling="2025-10-05 09:15:02.857151174 +0000 UTC m=+8536.739777634" lastFinishedPulling="2025-10-05 09:15:08.541788441 +0000 UTC m=+8542.424414911" observedRunningTime="2025-10-05 09:15:08.95758833 +0000 UTC m=+8542.840214830" watchObservedRunningTime="2025-10-05 09:15:08.958773951 +0000 UTC m=+8542.841400421" Oct 05 09:15:11 crc kubenswrapper[4935]: I1005 09:15:11.338293 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:11 crc kubenswrapper[4935]: I1005 09:15:11.338783 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:12 crc kubenswrapper[4935]: I1005 09:15:12.391273 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p2d2q" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="registry-server" probeResult="failure" output=< Oct 05 09:15:12 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 09:15:12 crc kubenswrapper[4935]: > Oct 05 09:15:14 crc kubenswrapper[4935]: I1005 09:15:14.777250 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:15:14 crc kubenswrapper[4935]: E1005 09:15:14.777714 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:15:21 crc kubenswrapper[4935]: I1005 09:15:21.409992 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:21 crc kubenswrapper[4935]: I1005 09:15:21.460617 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:21 crc kubenswrapper[4935]: I1005 09:15:21.644588 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p2d2q"] Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.083328 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p2d2q" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="registry-server" containerID="cri-o://f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524" gracePeriod=2 Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.728811 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.846952 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-utilities\") pod \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.847368 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p67ms\" (UniqueName: \"kubernetes.io/projected/be733bf3-217f-4f6d-8a22-5ab12b4c2528-kube-api-access-p67ms\") pod \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.847578 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-catalog-content\") pod \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\" (UID: \"be733bf3-217f-4f6d-8a22-5ab12b4c2528\") " Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.852312 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-utilities" (OuterVolumeSpecName: "utilities") pod "be733bf3-217f-4f6d-8a22-5ab12b4c2528" (UID: "be733bf3-217f-4f6d-8a22-5ab12b4c2528"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.872219 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be733bf3-217f-4f6d-8a22-5ab12b4c2528-kube-api-access-p67ms" (OuterVolumeSpecName: "kube-api-access-p67ms") pod "be733bf3-217f-4f6d-8a22-5ab12b4c2528" (UID: "be733bf3-217f-4f6d-8a22-5ab12b4c2528"). InnerVolumeSpecName "kube-api-access-p67ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.956109 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.956146 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p67ms\" (UniqueName: \"kubernetes.io/projected/be733bf3-217f-4f6d-8a22-5ab12b4c2528-kube-api-access-p67ms\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:23 crc kubenswrapper[4935]: I1005 09:15:23.956469 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be733bf3-217f-4f6d-8a22-5ab12b4c2528" (UID: "be733bf3-217f-4f6d-8a22-5ab12b4c2528"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.057473 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be733bf3-217f-4f6d-8a22-5ab12b4c2528-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.105385 4935 generic.go:334] "Generic (PLEG): container finished" podID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerID="f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524" exitCode=0 Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.105451 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2d2q" event={"ID":"be733bf3-217f-4f6d-8a22-5ab12b4c2528","Type":"ContainerDied","Data":"f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524"} Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.105481 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2d2q" event={"ID":"be733bf3-217f-4f6d-8a22-5ab12b4c2528","Type":"ContainerDied","Data":"c1ad823677aead39424c6a1459a847b4b609d450385562bac0538071e364e0c8"} Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.105500 4935 scope.go:117] "RemoveContainer" containerID="f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.105721 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p2d2q" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.143375 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p2d2q"] Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.152599 4935 scope.go:117] "RemoveContainer" containerID="eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.155407 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p2d2q"] Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.178387 4935 scope.go:117] "RemoveContainer" containerID="119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.216092 4935 scope.go:117] "RemoveContainer" containerID="f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524" Oct 05 09:15:24 crc kubenswrapper[4935]: E1005 09:15:24.216542 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524\": container with ID starting with f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524 not found: ID does not exist" containerID="f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.216616 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524"} err="failed to get container status \"f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524\": rpc error: code = NotFound desc = could not find container \"f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524\": container with ID starting with f852e7bdbec94c2a4c99b28f3275249ec208ef5162e25ced5a88dc1a24337524 not found: ID does not exist" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.216642 4935 scope.go:117] "RemoveContainer" containerID="eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62" Oct 05 09:15:24 crc kubenswrapper[4935]: E1005 09:15:24.217543 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62\": container with ID starting with eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62 not found: ID does not exist" containerID="eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.217664 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62"} err="failed to get container status \"eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62\": rpc error: code = NotFound desc = could not find container \"eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62\": container with ID starting with eeab20ac025f24655434e28cf1f1bcc03cc60eb335c107ba925060d0d2addc62 not found: ID does not exist" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.217714 4935 scope.go:117] "RemoveContainer" containerID="119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b" Oct 05 09:15:24 crc kubenswrapper[4935]: E1005 09:15:24.218313 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b\": container with ID starting with 119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b not found: ID does not exist" containerID="119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.218366 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b"} err="failed to get container status \"119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b\": rpc error: code = NotFound desc = could not find container \"119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b\": container with ID starting with 119118c0c859078da52b702df75e65c74c127616d39514c42c9fd74e5e16dc4b not found: ID does not exist" Oct 05 09:15:24 crc kubenswrapper[4935]: I1005 09:15:24.796119 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" path="/var/lib/kubelet/pods/be733bf3-217f-4f6d-8a22-5ab12b4c2528/volumes" Oct 05 09:15:25 crc kubenswrapper[4935]: I1005 09:15:25.778694 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:15:25 crc kubenswrapper[4935]: E1005 09:15:25.779225 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:15:40 crc kubenswrapper[4935]: I1005 09:15:40.777814 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:15:40 crc kubenswrapper[4935]: E1005 09:15:40.778996 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:15:55 crc kubenswrapper[4935]: I1005 09:15:55.777434 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:15:55 crc kubenswrapper[4935]: E1005 09:15:55.778231 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:16:03 crc kubenswrapper[4935]: I1005 09:16:03.217913 4935 scope.go:117] "RemoveContainer" containerID="0f7ff305a89f1eae0a5a300a0eeda823263feaf99b6ef15ed6d853f06474b825" Oct 05 09:16:06 crc kubenswrapper[4935]: I1005 09:16:06.785525 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:16:06 crc kubenswrapper[4935]: E1005 09:16:06.786829 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:16:12 crc kubenswrapper[4935]: I1005 09:16:12.611318 4935 generic.go:334] "Generic (PLEG): container finished" podID="6d1fcc45-8bfe-4406-ab6d-fa30a85985cc" containerID="8ef49d37538c2383d8c833fdf21a0736f23e632980d0ee0721670b862cc98628" exitCode=0 Oct 05 09:16:12 crc kubenswrapper[4935]: I1005 09:16:12.611455 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-blbmn" event={"ID":"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc","Type":"ContainerDied","Data":"8ef49d37538c2383d8c833fdf21a0736f23e632980d0ee0721670b862cc98628"} Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.206137 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.290607 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-ssh-key\") pod \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.290654 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-inventory\") pod \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.290675 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skjbw\" (UniqueName: \"kubernetes.io/projected/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-kube-api-access-skjbw\") pod \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\" (UID: \"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc\") " Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.297991 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-kube-api-access-skjbw" (OuterVolumeSpecName: "kube-api-access-skjbw") pod "6d1fcc45-8bfe-4406-ab6d-fa30a85985cc" (UID: "6d1fcc45-8bfe-4406-ab6d-fa30a85985cc"). InnerVolumeSpecName "kube-api-access-skjbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.319562 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-inventory" (OuterVolumeSpecName: "inventory") pod "6d1fcc45-8bfe-4406-ab6d-fa30a85985cc" (UID: "6d1fcc45-8bfe-4406-ab6d-fa30a85985cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.354096 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6d1fcc45-8bfe-4406-ab6d-fa30a85985cc" (UID: "6d1fcc45-8bfe-4406-ab6d-fa30a85985cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.392828 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.392869 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.392883 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skjbw\" (UniqueName: \"kubernetes.io/projected/6d1fcc45-8bfe-4406-ab6d-fa30a85985cc-kube-api-access-skjbw\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.635287 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-blbmn" event={"ID":"6d1fcc45-8bfe-4406-ab6d-fa30a85985cc","Type":"ContainerDied","Data":"4a78e411fd7dac0ca03d8f2b3ca6a0a121f09b27a683d7e264af3f2d20628acb"} Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.635333 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a78e411fd7dac0ca03d8f2b3ca6a0a121f09b27a683d7e264af3f2d20628acb" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.635345 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-blbmn" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.767371 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-networker-wq6jl"] Oct 05 09:16:14 crc kubenswrapper[4935]: E1005 09:16:14.767769 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="extract-utilities" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.767785 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="extract-utilities" Oct 05 09:16:14 crc kubenswrapper[4935]: E1005 09:16:14.767793 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="registry-server" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.767801 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="registry-server" Oct 05 09:16:14 crc kubenswrapper[4935]: E1005 09:16:14.767822 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1fcc45-8bfe-4406-ab6d-fa30a85985cc" containerName="download-cache-openstack-openstack-networker" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.767830 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1fcc45-8bfe-4406-ab6d-fa30a85985cc" containerName="download-cache-openstack-openstack-networker" Oct 05 09:16:14 crc kubenswrapper[4935]: E1005 09:16:14.767846 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="extract-content" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.767853 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="extract-content" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.768046 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="be733bf3-217f-4f6d-8a22-5ab12b4c2528" containerName="registry-server" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.768068 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1fcc45-8bfe-4406-ab6d-fa30a85985cc" containerName="download-cache-openstack-openstack-networker" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.769214 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.796050 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.796316 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.824149 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-ssh-key\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.824536 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c7lk\" (UniqueName: \"kubernetes.io/projected/2bec0abb-c7e2-4a0b-8075-83158bcd560f-kube-api-access-5c7lk\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.824801 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-inventory\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.879131 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-wq6jl"] Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.943175 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-ssh-key\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.943231 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c7lk\" (UniqueName: \"kubernetes.io/projected/2bec0abb-c7e2-4a0b-8075-83158bcd560f-kube-api-access-5c7lk\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.943280 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-inventory\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.954655 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-ssh-key\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.957830 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-inventory\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:14 crc kubenswrapper[4935]: I1005 09:16:14.968575 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c7lk\" (UniqueName: \"kubernetes.io/projected/2bec0abb-c7e2-4a0b-8075-83158bcd560f-kube-api-access-5c7lk\") pod \"configure-network-openstack-openstack-networker-wq6jl\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:15 crc kubenswrapper[4935]: I1005 09:16:15.096687 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:16:15 crc kubenswrapper[4935]: I1005 09:16:15.684866 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-wq6jl"] Oct 05 09:16:16 crc kubenswrapper[4935]: I1005 09:16:16.658003 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" event={"ID":"2bec0abb-c7e2-4a0b-8075-83158bcd560f","Type":"ContainerStarted","Data":"aff67d29a4d20dda1aa5e97d84ff29d9df7fa646ad9fda2cf6f88535687e519b"} Oct 05 09:16:16 crc kubenswrapper[4935]: I1005 09:16:16.658403 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" event={"ID":"2bec0abb-c7e2-4a0b-8075-83158bcd560f","Type":"ContainerStarted","Data":"5668d1e9632e1331895168f733748bc872344d43c07f3bff5c77787f9fcd5c5f"} Oct 05 09:16:16 crc kubenswrapper[4935]: I1005 09:16:16.683314 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" podStartSLOduration=2.052679948 podStartE2EDuration="2.683295622s" podCreationTimestamp="2025-10-05 09:16:14 +0000 UTC" firstStartedPulling="2025-10-05 09:16:15.679773457 +0000 UTC m=+8609.562399907" lastFinishedPulling="2025-10-05 09:16:16.310389091 +0000 UTC m=+8610.193015581" observedRunningTime="2025-10-05 09:16:16.681538266 +0000 UTC m=+8610.564164726" watchObservedRunningTime="2025-10-05 09:16:16.683295622 +0000 UTC m=+8610.565922082" Oct 05 09:16:18 crc kubenswrapper[4935]: I1005 09:16:18.778591 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:16:18 crc kubenswrapper[4935]: E1005 09:16:18.779295 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:16:26 crc kubenswrapper[4935]: I1005 09:16:26.785713 4935 generic.go:334] "Generic (PLEG): container finished" podID="e39fcf7d-8ea6-4a0a-87ed-a6000793d030" containerID="d484ae8f39fe144b8817f23815ae50472d12f079a7102606bbb3138d72cf33f8" exitCode=0 Oct 05 09:16:26 crc kubenswrapper[4935]: I1005 09:16:26.800198 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" event={"ID":"e39fcf7d-8ea6-4a0a-87ed-a6000793d030","Type":"ContainerDied","Data":"d484ae8f39fe144b8817f23815ae50472d12f079a7102606bbb3138d72cf33f8"} Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.301230 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.359200 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ssh-key\") pod \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.359457 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ceph\") pod \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.359490 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk2r2\" (UniqueName: \"kubernetes.io/projected/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-kube-api-access-gk2r2\") pod \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.359617 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-inventory\") pod \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\" (UID: \"e39fcf7d-8ea6-4a0a-87ed-a6000793d030\") " Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.365411 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ceph" (OuterVolumeSpecName: "ceph") pod "e39fcf7d-8ea6-4a0a-87ed-a6000793d030" (UID: "e39fcf7d-8ea6-4a0a-87ed-a6000793d030"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.365578 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-kube-api-access-gk2r2" (OuterVolumeSpecName: "kube-api-access-gk2r2") pod "e39fcf7d-8ea6-4a0a-87ed-a6000793d030" (UID: "e39fcf7d-8ea6-4a0a-87ed-a6000793d030"). InnerVolumeSpecName "kube-api-access-gk2r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.388108 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-inventory" (OuterVolumeSpecName: "inventory") pod "e39fcf7d-8ea6-4a0a-87ed-a6000793d030" (UID: "e39fcf7d-8ea6-4a0a-87ed-a6000793d030"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.389389 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e39fcf7d-8ea6-4a0a-87ed-a6000793d030" (UID: "e39fcf7d-8ea6-4a0a-87ed-a6000793d030"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.461743 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.461781 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk2r2\" (UniqueName: \"kubernetes.io/projected/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-kube-api-access-gk2r2\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.461794 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.461803 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e39fcf7d-8ea6-4a0a-87ed-a6000793d030-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.815534 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" event={"ID":"e39fcf7d-8ea6-4a0a-87ed-a6000793d030","Type":"ContainerDied","Data":"78f79666665b36ff35fcd9945a8dd179aa1223fc898f7efcf8517bf6576964d7"} Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.815604 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f79666665b36ff35fcd9945a8dd179aa1223fc898f7efcf8517bf6576964d7" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.815736 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-ssp86" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.924422 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-6j8z7"] Oct 05 09:16:28 crc kubenswrapper[4935]: E1005 09:16:28.924990 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39fcf7d-8ea6-4a0a-87ed-a6000793d030" containerName="download-cache-openstack-openstack-cell1" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.925016 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39fcf7d-8ea6-4a0a-87ed-a6000793d030" containerName="download-cache-openstack-openstack-cell1" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.925292 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e39fcf7d-8ea6-4a0a-87ed-a6000793d030" containerName="download-cache-openstack-openstack-cell1" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.926338 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.929909 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.931211 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.939618 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-6j8z7"] Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.973189 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ceph\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.973525 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ssh-key\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.973663 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-inventory\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:28 crc kubenswrapper[4935]: I1005 09:16:28.973787 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jsfx\" (UniqueName: \"kubernetes.io/projected/a90bd676-a895-44f5-b3b5-c23a9178c9e0-kube-api-access-5jsfx\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.075435 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ssh-key\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.075505 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-inventory\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.075567 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jsfx\" (UniqueName: \"kubernetes.io/projected/a90bd676-a895-44f5-b3b5-c23a9178c9e0-kube-api-access-5jsfx\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.075618 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ceph\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.080223 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ssh-key\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.080524 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-inventory\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.081679 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ceph\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.099480 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jsfx\" (UniqueName: \"kubernetes.io/projected/a90bd676-a895-44f5-b3b5-c23a9178c9e0-kube-api-access-5jsfx\") pod \"configure-network-openstack-openstack-cell1-6j8z7\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.257683 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.780258 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:16:29 crc kubenswrapper[4935]: E1005 09:16:29.780585 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:16:29 crc kubenswrapper[4935]: I1005 09:16:29.835783 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-6j8z7"] Oct 05 09:16:29 crc kubenswrapper[4935]: W1005 09:16:29.836265 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda90bd676_a895_44f5_b3b5_c23a9178c9e0.slice/crio-d515849b01e1e0184634c5bcb5f8de9ad064730c5ec4cf0557bf1b2e91a2364d WatchSource:0}: Error finding container d515849b01e1e0184634c5bcb5f8de9ad064730c5ec4cf0557bf1b2e91a2364d: Status 404 returned error can't find the container with id d515849b01e1e0184634c5bcb5f8de9ad064730c5ec4cf0557bf1b2e91a2364d Oct 05 09:16:30 crc kubenswrapper[4935]: I1005 09:16:30.841687 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" event={"ID":"a90bd676-a895-44f5-b3b5-c23a9178c9e0","Type":"ContainerStarted","Data":"92f21263a3cc410f7557fcbe94832f7fccbe1f1a0b019e154f1394a2634537cc"} Oct 05 09:16:30 crc kubenswrapper[4935]: I1005 09:16:30.842332 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" event={"ID":"a90bd676-a895-44f5-b3b5-c23a9178c9e0","Type":"ContainerStarted","Data":"d515849b01e1e0184634c5bcb5f8de9ad064730c5ec4cf0557bf1b2e91a2364d"} Oct 05 09:16:30 crc kubenswrapper[4935]: I1005 09:16:30.871681 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" podStartSLOduration=2.460929075 podStartE2EDuration="2.87166062s" podCreationTimestamp="2025-10-05 09:16:28 +0000 UTC" firstStartedPulling="2025-10-05 09:16:29.839733208 +0000 UTC m=+8623.722359668" lastFinishedPulling="2025-10-05 09:16:30.250464753 +0000 UTC m=+8624.133091213" observedRunningTime="2025-10-05 09:16:30.855107005 +0000 UTC m=+8624.737733485" watchObservedRunningTime="2025-10-05 09:16:30.87166062 +0000 UTC m=+8624.754287090" Oct 05 09:16:40 crc kubenswrapper[4935]: I1005 09:16:40.778086 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:16:40 crc kubenswrapper[4935]: E1005 09:16:40.779124 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:16:52 crc kubenswrapper[4935]: I1005 09:16:52.777238 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:16:52 crc kubenswrapper[4935]: E1005 09:16:52.778189 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:17:04 crc kubenswrapper[4935]: I1005 09:17:04.777615 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:17:04 crc kubenswrapper[4935]: E1005 09:17:04.778652 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:17:17 crc kubenswrapper[4935]: I1005 09:17:17.778174 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:17:18 crc kubenswrapper[4935]: I1005 09:17:18.371726 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"989e4812ce9b8760f616df4a134bbde12de0cc616591fdb55b6764cc85626afe"} Oct 05 09:17:21 crc kubenswrapper[4935]: I1005 09:17:21.413669 4935 generic.go:334] "Generic (PLEG): container finished" podID="2bec0abb-c7e2-4a0b-8075-83158bcd560f" containerID="aff67d29a4d20dda1aa5e97d84ff29d9df7fa646ad9fda2cf6f88535687e519b" exitCode=0 Oct 05 09:17:21 crc kubenswrapper[4935]: I1005 09:17:21.414259 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" event={"ID":"2bec0abb-c7e2-4a0b-8075-83158bcd560f","Type":"ContainerDied","Data":"aff67d29a4d20dda1aa5e97d84ff29d9df7fa646ad9fda2cf6f88535687e519b"} Oct 05 09:17:22 crc kubenswrapper[4935]: I1005 09:17:22.968409 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.064929 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-ssh-key\") pod \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.065137 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c7lk\" (UniqueName: \"kubernetes.io/projected/2bec0abb-c7e2-4a0b-8075-83158bcd560f-kube-api-access-5c7lk\") pod \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.065422 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-inventory\") pod \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\" (UID: \"2bec0abb-c7e2-4a0b-8075-83158bcd560f\") " Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.070588 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bec0abb-c7e2-4a0b-8075-83158bcd560f-kube-api-access-5c7lk" (OuterVolumeSpecName: "kube-api-access-5c7lk") pod "2bec0abb-c7e2-4a0b-8075-83158bcd560f" (UID: "2bec0abb-c7e2-4a0b-8075-83158bcd560f"). InnerVolumeSpecName "kube-api-access-5c7lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.105955 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2bec0abb-c7e2-4a0b-8075-83158bcd560f" (UID: "2bec0abb-c7e2-4a0b-8075-83158bcd560f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.119068 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-inventory" (OuterVolumeSpecName: "inventory") pod "2bec0abb-c7e2-4a0b-8075-83158bcd560f" (UID: "2bec0abb-c7e2-4a0b-8075-83158bcd560f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.167986 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.168036 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c7lk\" (UniqueName: \"kubernetes.io/projected/2bec0abb-c7e2-4a0b-8075-83158bcd560f-kube-api-access-5c7lk\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.168052 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bec0abb-c7e2-4a0b-8075-83158bcd560f-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.432036 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" event={"ID":"2bec0abb-c7e2-4a0b-8075-83158bcd560f","Type":"ContainerDied","Data":"5668d1e9632e1331895168f733748bc872344d43c07f3bff5c77787f9fcd5c5f"} Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.432091 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-wq6jl" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.432107 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5668d1e9632e1331895168f733748bc872344d43c07f3bff5c77787f9fcd5c5f" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.522563 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-networker-cchlw"] Oct 05 09:17:23 crc kubenswrapper[4935]: E1005 09:17:23.523117 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bec0abb-c7e2-4a0b-8075-83158bcd560f" containerName="configure-network-openstack-openstack-networker" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.523163 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bec0abb-c7e2-4a0b-8075-83158bcd560f" containerName="configure-network-openstack-openstack-networker" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.523588 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bec0abb-c7e2-4a0b-8075-83158bcd560f" containerName="configure-network-openstack-openstack-networker" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.524539 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.526857 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.527096 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.536417 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-cchlw"] Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.677640 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-inventory\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.677702 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-ssh-key\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.677830 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfh9v\" (UniqueName: \"kubernetes.io/projected/9ea0d044-3609-43f6-af64-15f5901287ad-kube-api-access-lfh9v\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.779962 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-inventory\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.780338 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-ssh-key\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.780380 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfh9v\" (UniqueName: \"kubernetes.io/projected/9ea0d044-3609-43f6-af64-15f5901287ad-kube-api-access-lfh9v\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.785202 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-ssh-key\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.786079 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-inventory\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.795231 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfh9v\" (UniqueName: \"kubernetes.io/projected/9ea0d044-3609-43f6-af64-15f5901287ad-kube-api-access-lfh9v\") pod \"validate-network-openstack-openstack-networker-cchlw\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:23 crc kubenswrapper[4935]: I1005 09:17:23.841252 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:24 crc kubenswrapper[4935]: I1005 09:17:24.398951 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-cchlw"] Oct 05 09:17:24 crc kubenswrapper[4935]: W1005 09:17:24.406434 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ea0d044_3609_43f6_af64_15f5901287ad.slice/crio-ac84c8f2fe6a3ed12ffa1dae35168f51ea0c8214d2d19b4e6cbc85f4b7c68135 WatchSource:0}: Error finding container ac84c8f2fe6a3ed12ffa1dae35168f51ea0c8214d2d19b4e6cbc85f4b7c68135: Status 404 returned error can't find the container with id ac84c8f2fe6a3ed12ffa1dae35168f51ea0c8214d2d19b4e6cbc85f4b7c68135 Oct 05 09:17:24 crc kubenswrapper[4935]: I1005 09:17:24.444245 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-cchlw" event={"ID":"9ea0d044-3609-43f6-af64-15f5901287ad","Type":"ContainerStarted","Data":"ac84c8f2fe6a3ed12ffa1dae35168f51ea0c8214d2d19b4e6cbc85f4b7c68135"} Oct 05 09:17:25 crc kubenswrapper[4935]: I1005 09:17:25.457610 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-cchlw" event={"ID":"9ea0d044-3609-43f6-af64-15f5901287ad","Type":"ContainerStarted","Data":"ec6db67b848c4af403b2b12d44206feb93630e05becb5dcfc086265af486eb38"} Oct 05 09:17:25 crc kubenswrapper[4935]: I1005 09:17:25.476440 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-networker-cchlw" podStartSLOduration=1.864417172 podStartE2EDuration="2.476412567s" podCreationTimestamp="2025-10-05 09:17:23 +0000 UTC" firstStartedPulling="2025-10-05 09:17:24.409017763 +0000 UTC m=+8678.291644223" lastFinishedPulling="2025-10-05 09:17:25.021013158 +0000 UTC m=+8678.903639618" observedRunningTime="2025-10-05 09:17:25.470454221 +0000 UTC m=+8679.353080691" watchObservedRunningTime="2025-10-05 09:17:25.476412567 +0000 UTC m=+8679.359039057" Oct 05 09:17:30 crc kubenswrapper[4935]: I1005 09:17:30.507560 4935 generic.go:334] "Generic (PLEG): container finished" podID="9ea0d044-3609-43f6-af64-15f5901287ad" containerID="ec6db67b848c4af403b2b12d44206feb93630e05becb5dcfc086265af486eb38" exitCode=0 Oct 05 09:17:30 crc kubenswrapper[4935]: I1005 09:17:30.507684 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-cchlw" event={"ID":"9ea0d044-3609-43f6-af64-15f5901287ad","Type":"ContainerDied","Data":"ec6db67b848c4af403b2b12d44206feb93630e05becb5dcfc086265af486eb38"} Oct 05 09:17:31 crc kubenswrapper[4935]: I1005 09:17:31.519351 4935 generic.go:334] "Generic (PLEG): container finished" podID="a90bd676-a895-44f5-b3b5-c23a9178c9e0" containerID="92f21263a3cc410f7557fcbe94832f7fccbe1f1a0b019e154f1394a2634537cc" exitCode=0 Oct 05 09:17:31 crc kubenswrapper[4935]: I1005 09:17:31.519432 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" event={"ID":"a90bd676-a895-44f5-b3b5-c23a9178c9e0","Type":"ContainerDied","Data":"92f21263a3cc410f7557fcbe94832f7fccbe1f1a0b019e154f1394a2634537cc"} Oct 05 09:17:31 crc kubenswrapper[4935]: I1005 09:17:31.960322 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.063825 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-ssh-key\") pod \"9ea0d044-3609-43f6-af64-15f5901287ad\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.064047 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfh9v\" (UniqueName: \"kubernetes.io/projected/9ea0d044-3609-43f6-af64-15f5901287ad-kube-api-access-lfh9v\") pod \"9ea0d044-3609-43f6-af64-15f5901287ad\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.064220 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-inventory\") pod \"9ea0d044-3609-43f6-af64-15f5901287ad\" (UID: \"9ea0d044-3609-43f6-af64-15f5901287ad\") " Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.071043 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ea0d044-3609-43f6-af64-15f5901287ad-kube-api-access-lfh9v" (OuterVolumeSpecName: "kube-api-access-lfh9v") pod "9ea0d044-3609-43f6-af64-15f5901287ad" (UID: "9ea0d044-3609-43f6-af64-15f5901287ad"). InnerVolumeSpecName "kube-api-access-lfh9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.103378 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9ea0d044-3609-43f6-af64-15f5901287ad" (UID: "9ea0d044-3609-43f6-af64-15f5901287ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.110148 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-inventory" (OuterVolumeSpecName: "inventory") pod "9ea0d044-3609-43f6-af64-15f5901287ad" (UID: "9ea0d044-3609-43f6-af64-15f5901287ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.167230 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.167527 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfh9v\" (UniqueName: \"kubernetes.io/projected/9ea0d044-3609-43f6-af64-15f5901287ad-kube-api-access-lfh9v\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.167538 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ea0d044-3609-43f6-af64-15f5901287ad-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.529646 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-cchlw" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.529643 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-cchlw" event={"ID":"9ea0d044-3609-43f6-af64-15f5901287ad","Type":"ContainerDied","Data":"ac84c8f2fe6a3ed12ffa1dae35168f51ea0c8214d2d19b4e6cbc85f4b7c68135"} Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.529923 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac84c8f2fe6a3ed12ffa1dae35168f51ea0c8214d2d19b4e6cbc85f4b7c68135" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.716388 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-networker-xkr98"] Oct 05 09:17:32 crc kubenswrapper[4935]: E1005 09:17:32.717008 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ea0d044-3609-43f6-af64-15f5901287ad" containerName="validate-network-openstack-openstack-networker" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.717028 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ea0d044-3609-43f6-af64-15f5901287ad" containerName="validate-network-openstack-openstack-networker" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.729118 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ea0d044-3609-43f6-af64-15f5901287ad" containerName="validate-network-openstack-openstack-networker" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.730582 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-xkr98"] Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.730703 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.746034 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.746271 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.887759 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-inventory\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.887870 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-ssh-key\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.888690 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s268\" (UniqueName: \"kubernetes.io/projected/cdbef832-881e-4a2a-8735-59f801251dd5-kube-api-access-6s268\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.991292 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s268\" (UniqueName: \"kubernetes.io/projected/cdbef832-881e-4a2a-8735-59f801251dd5-kube-api-access-6s268\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.992121 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-inventory\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.992227 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-ssh-key\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.997952 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-inventory\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:32 crc kubenswrapper[4935]: I1005 09:17:32.998402 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-ssh-key\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.011499 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s268\" (UniqueName: \"kubernetes.io/projected/cdbef832-881e-4a2a-8735-59f801251dd5-kube-api-access-6s268\") pod \"install-os-openstack-openstack-networker-xkr98\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.060687 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.070771 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.195863 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jsfx\" (UniqueName: \"kubernetes.io/projected/a90bd676-a895-44f5-b3b5-c23a9178c9e0-kube-api-access-5jsfx\") pod \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.196035 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ssh-key\") pod \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.196097 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-inventory\") pod \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.196333 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ceph\") pod \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\" (UID: \"a90bd676-a895-44f5-b3b5-c23a9178c9e0\") " Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.200042 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90bd676-a895-44f5-b3b5-c23a9178c9e0-kube-api-access-5jsfx" (OuterVolumeSpecName: "kube-api-access-5jsfx") pod "a90bd676-a895-44f5-b3b5-c23a9178c9e0" (UID: "a90bd676-a895-44f5-b3b5-c23a9178c9e0"). InnerVolumeSpecName "kube-api-access-5jsfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.202060 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ceph" (OuterVolumeSpecName: "ceph") pod "a90bd676-a895-44f5-b3b5-c23a9178c9e0" (UID: "a90bd676-a895-44f5-b3b5-c23a9178c9e0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.243075 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-inventory" (OuterVolumeSpecName: "inventory") pod "a90bd676-a895-44f5-b3b5-c23a9178c9e0" (UID: "a90bd676-a895-44f5-b3b5-c23a9178c9e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.245174 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a90bd676-a895-44f5-b3b5-c23a9178c9e0" (UID: "a90bd676-a895-44f5-b3b5-c23a9178c9e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.301600 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.301638 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jsfx\" (UniqueName: \"kubernetes.io/projected/a90bd676-a895-44f5-b3b5-c23a9178c9e0-kube-api-access-5jsfx\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.301653 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.301665 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90bd676-a895-44f5-b3b5-c23a9178c9e0-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.430453 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-xkr98"] Oct 05 09:17:33 crc kubenswrapper[4935]: W1005 09:17:33.433828 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdbef832_881e_4a2a_8735_59f801251dd5.slice/crio-1745353a8ca0068f7e1bc8367d92501d9910069c0a71121da5a7b3c9235bf10a WatchSource:0}: Error finding container 1745353a8ca0068f7e1bc8367d92501d9910069c0a71121da5a7b3c9235bf10a: Status 404 returned error can't find the container with id 1745353a8ca0068f7e1bc8367d92501d9910069c0a71121da5a7b3c9235bf10a Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.539081 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.539071 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-6j8z7" event={"ID":"a90bd676-a895-44f5-b3b5-c23a9178c9e0","Type":"ContainerDied","Data":"d515849b01e1e0184634c5bcb5f8de9ad064730c5ec4cf0557bf1b2e91a2364d"} Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.539256 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d515849b01e1e0184634c5bcb5f8de9ad064730c5ec4cf0557bf1b2e91a2364d" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.540601 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-xkr98" event={"ID":"cdbef832-881e-4a2a-8735-59f801251dd5","Type":"ContainerStarted","Data":"1745353a8ca0068f7e1bc8367d92501d9910069c0a71121da5a7b3c9235bf10a"} Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.612860 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-6nl7k"] Oct 05 09:17:33 crc kubenswrapper[4935]: E1005 09:17:33.613437 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90bd676-a895-44f5-b3b5-c23a9178c9e0" containerName="configure-network-openstack-openstack-cell1" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.613460 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90bd676-a895-44f5-b3b5-c23a9178c9e0" containerName="configure-network-openstack-openstack-cell1" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.613754 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a90bd676-a895-44f5-b3b5-c23a9178c9e0" containerName="configure-network-openstack-openstack-cell1" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.614713 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.617808 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.618175 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.631478 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-6nl7k"] Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.709013 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-inventory\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.709133 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ssh-key\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.709215 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ceph\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.709283 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85cxz\" (UniqueName: \"kubernetes.io/projected/ed429030-8625-49b1-8e3f-106b40ae136a-kube-api-access-85cxz\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.810958 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ceph\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.811039 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85cxz\" (UniqueName: \"kubernetes.io/projected/ed429030-8625-49b1-8e3f-106b40ae136a-kube-api-access-85cxz\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.811150 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-inventory\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.811234 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ssh-key\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.814678 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-inventory\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.815478 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ceph\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.815842 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ssh-key\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.826844 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85cxz\" (UniqueName: \"kubernetes.io/projected/ed429030-8625-49b1-8e3f-106b40ae136a-kube-api-access-85cxz\") pod \"validate-network-openstack-openstack-cell1-6nl7k\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:33 crc kubenswrapper[4935]: I1005 09:17:33.933120 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.127336 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mm6wt"] Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.130307 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.165475 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mm6wt"] Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.223635 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-utilities\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.223737 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wp9v\" (UniqueName: \"kubernetes.io/projected/6b6e78d1-72d5-44de-9d15-827b2303e6a0-kube-api-access-9wp9v\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.223877 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-catalog-content\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.326406 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-catalog-content\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.327042 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-utilities\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.327106 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wp9v\" (UniqueName: \"kubernetes.io/projected/6b6e78d1-72d5-44de-9d15-827b2303e6a0-kube-api-access-9wp9v\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.327226 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-catalog-content\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.329852 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-utilities\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.345071 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wp9v\" (UniqueName: \"kubernetes.io/projected/6b6e78d1-72d5-44de-9d15-827b2303e6a0-kube-api-access-9wp9v\") pod \"certified-operators-mm6wt\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.472386 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.556059 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-xkr98" event={"ID":"cdbef832-881e-4a2a-8735-59f801251dd5","Type":"ContainerStarted","Data":"7c29d2c0a0459c3a95c3c3eb5c2c6b31bccfc0220a7f85e303a4628c21b84d91"} Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.593818 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-networker-xkr98" podStartSLOduration=1.991939736 podStartE2EDuration="2.593802264s" podCreationTimestamp="2025-10-05 09:17:32 +0000 UTC" firstStartedPulling="2025-10-05 09:17:33.436567889 +0000 UTC m=+8687.319194339" lastFinishedPulling="2025-10-05 09:17:34.038430367 +0000 UTC m=+8687.921056867" observedRunningTime="2025-10-05 09:17:34.587338844 +0000 UTC m=+8688.469965314" watchObservedRunningTime="2025-10-05 09:17:34.593802264 +0000 UTC m=+8688.476428714" Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.629356 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-6nl7k"] Oct 05 09:17:34 crc kubenswrapper[4935]: W1005 09:17:34.985306 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b6e78d1_72d5_44de_9d15_827b2303e6a0.slice/crio-6d75e078e0d0580c40c4bf66f7b46fdbd5dc682f14c9fa7e536772b03dc9a3cc WatchSource:0}: Error finding container 6d75e078e0d0580c40c4bf66f7b46fdbd5dc682f14c9fa7e536772b03dc9a3cc: Status 404 returned error can't find the container with id 6d75e078e0d0580c40c4bf66f7b46fdbd5dc682f14c9fa7e536772b03dc9a3cc Oct 05 09:17:34 crc kubenswrapper[4935]: I1005 09:17:34.998655 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mm6wt"] Oct 05 09:17:35 crc kubenswrapper[4935]: E1005 09:17:35.353009 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b6e78d1_72d5_44de_9d15_827b2303e6a0.slice/crio-conmon-cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2.scope\": RecentStats: unable to find data in memory cache]" Oct 05 09:17:35 crc kubenswrapper[4935]: I1005 09:17:35.566427 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" event={"ID":"ed429030-8625-49b1-8e3f-106b40ae136a","Type":"ContainerStarted","Data":"cec2288cd1b2415aa192fe0c4d1106f2cfad42b91b9e3fc65ff22d6e59982ae3"} Oct 05 09:17:35 crc kubenswrapper[4935]: I1005 09:17:35.566776 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" event={"ID":"ed429030-8625-49b1-8e3f-106b40ae136a","Type":"ContainerStarted","Data":"67663b57276109e2257809ae34db5b97efba1ebc6b6550496d04b98d104c8c33"} Oct 05 09:17:35 crc kubenswrapper[4935]: I1005 09:17:35.568235 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerID="cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2" exitCode=0 Oct 05 09:17:35 crc kubenswrapper[4935]: I1005 09:17:35.568342 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm6wt" event={"ID":"6b6e78d1-72d5-44de-9d15-827b2303e6a0","Type":"ContainerDied","Data":"cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2"} Oct 05 09:17:35 crc kubenswrapper[4935]: I1005 09:17:35.568394 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm6wt" event={"ID":"6b6e78d1-72d5-44de-9d15-827b2303e6a0","Type":"ContainerStarted","Data":"6d75e078e0d0580c40c4bf66f7b46fdbd5dc682f14c9fa7e536772b03dc9a3cc"} Oct 05 09:17:35 crc kubenswrapper[4935]: I1005 09:17:35.613320 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" podStartSLOduration=2.1729141849999998 podStartE2EDuration="2.613302969s" podCreationTimestamp="2025-10-05 09:17:33 +0000 UTC" firstStartedPulling="2025-10-05 09:17:34.657361535 +0000 UTC m=+8688.539987995" lastFinishedPulling="2025-10-05 09:17:35.097750319 +0000 UTC m=+8688.980376779" observedRunningTime="2025-10-05 09:17:35.605635258 +0000 UTC m=+8689.488261738" watchObservedRunningTime="2025-10-05 09:17:35.613302969 +0000 UTC m=+8689.495929429" Oct 05 09:17:36 crc kubenswrapper[4935]: I1005 09:17:36.582787 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm6wt" event={"ID":"6b6e78d1-72d5-44de-9d15-827b2303e6a0","Type":"ContainerStarted","Data":"0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7"} Oct 05 09:17:37 crc kubenswrapper[4935]: I1005 09:17:37.601350 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerID="0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7" exitCode=0 Oct 05 09:17:37 crc kubenswrapper[4935]: I1005 09:17:37.601417 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm6wt" event={"ID":"6b6e78d1-72d5-44de-9d15-827b2303e6a0","Type":"ContainerDied","Data":"0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7"} Oct 05 09:17:38 crc kubenswrapper[4935]: I1005 09:17:38.620170 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm6wt" event={"ID":"6b6e78d1-72d5-44de-9d15-827b2303e6a0","Type":"ContainerStarted","Data":"0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3"} Oct 05 09:17:38 crc kubenswrapper[4935]: I1005 09:17:38.647150 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mm6wt" podStartSLOduration=2.18862558 podStartE2EDuration="4.647120826s" podCreationTimestamp="2025-10-05 09:17:34 +0000 UTC" firstStartedPulling="2025-10-05 09:17:35.59278724 +0000 UTC m=+8689.475413700" lastFinishedPulling="2025-10-05 09:17:38.051282486 +0000 UTC m=+8691.933908946" observedRunningTime="2025-10-05 09:17:38.642048943 +0000 UTC m=+8692.524675423" watchObservedRunningTime="2025-10-05 09:17:38.647120826 +0000 UTC m=+8692.529747316" Oct 05 09:17:41 crc kubenswrapper[4935]: I1005 09:17:41.669427 4935 generic.go:334] "Generic (PLEG): container finished" podID="ed429030-8625-49b1-8e3f-106b40ae136a" containerID="cec2288cd1b2415aa192fe0c4d1106f2cfad42b91b9e3fc65ff22d6e59982ae3" exitCode=0 Oct 05 09:17:41 crc kubenswrapper[4935]: I1005 09:17:41.669538 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" event={"ID":"ed429030-8625-49b1-8e3f-106b40ae136a","Type":"ContainerDied","Data":"cec2288cd1b2415aa192fe0c4d1106f2cfad42b91b9e3fc65ff22d6e59982ae3"} Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.326915 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.430593 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85cxz\" (UniqueName: \"kubernetes.io/projected/ed429030-8625-49b1-8e3f-106b40ae136a-kube-api-access-85cxz\") pod \"ed429030-8625-49b1-8e3f-106b40ae136a\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.430767 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ssh-key\") pod \"ed429030-8625-49b1-8e3f-106b40ae136a\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.430809 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ceph\") pod \"ed429030-8625-49b1-8e3f-106b40ae136a\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.430879 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-inventory\") pod \"ed429030-8625-49b1-8e3f-106b40ae136a\" (UID: \"ed429030-8625-49b1-8e3f-106b40ae136a\") " Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.436403 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed429030-8625-49b1-8e3f-106b40ae136a-kube-api-access-85cxz" (OuterVolumeSpecName: "kube-api-access-85cxz") pod "ed429030-8625-49b1-8e3f-106b40ae136a" (UID: "ed429030-8625-49b1-8e3f-106b40ae136a"). InnerVolumeSpecName "kube-api-access-85cxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.437530 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ceph" (OuterVolumeSpecName: "ceph") pod "ed429030-8625-49b1-8e3f-106b40ae136a" (UID: "ed429030-8625-49b1-8e3f-106b40ae136a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.468615 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-inventory" (OuterVolumeSpecName: "inventory") pod "ed429030-8625-49b1-8e3f-106b40ae136a" (UID: "ed429030-8625-49b1-8e3f-106b40ae136a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.470403 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ed429030-8625-49b1-8e3f-106b40ae136a" (UID: "ed429030-8625-49b1-8e3f-106b40ae136a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.533627 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.533674 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.533689 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed429030-8625-49b1-8e3f-106b40ae136a-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.533702 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85cxz\" (UniqueName: \"kubernetes.io/projected/ed429030-8625-49b1-8e3f-106b40ae136a-kube-api-access-85cxz\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.700705 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" event={"ID":"ed429030-8625-49b1-8e3f-106b40ae136a","Type":"ContainerDied","Data":"67663b57276109e2257809ae34db5b97efba1ebc6b6550496d04b98d104c8c33"} Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.700751 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67663b57276109e2257809ae34db5b97efba1ebc6b6550496d04b98d104c8c33" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.700792 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-6nl7k" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.796071 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-bqz75"] Oct 05 09:17:43 crc kubenswrapper[4935]: E1005 09:17:43.796829 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed429030-8625-49b1-8e3f-106b40ae136a" containerName="validate-network-openstack-openstack-cell1" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.796965 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed429030-8625-49b1-8e3f-106b40ae136a" containerName="validate-network-openstack-openstack-cell1" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.797305 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed429030-8625-49b1-8e3f-106b40ae136a" containerName="validate-network-openstack-openstack-cell1" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.798212 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.806476 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.806718 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.807850 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-bqz75"] Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.943680 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ssh-key\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.943787 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s92p\" (UniqueName: \"kubernetes.io/projected/9113dc85-b57c-447f-ab4d-1294dcd0f61f-kube-api-access-6s92p\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.943931 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ceph\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:43 crc kubenswrapper[4935]: I1005 09:17:43.944203 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-inventory\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.046365 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-inventory\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.046810 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ssh-key\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.046872 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s92p\" (UniqueName: \"kubernetes.io/projected/9113dc85-b57c-447f-ab4d-1294dcd0f61f-kube-api-access-6s92p\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.047025 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ceph\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.053154 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ssh-key\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.053677 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-inventory\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.053820 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ceph\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.068429 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s92p\" (UniqueName: \"kubernetes.io/projected/9113dc85-b57c-447f-ab4d-1294dcd0f61f-kube-api-access-6s92p\") pod \"install-os-openstack-openstack-cell1-bqz75\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.124390 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.472995 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.473403 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.528821 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.695823 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-bqz75"] Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.767105 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:44 crc kubenswrapper[4935]: I1005 09:17:44.816424 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mm6wt"] Oct 05 09:17:45 crc kubenswrapper[4935]: I1005 09:17:45.719707 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-bqz75" event={"ID":"9113dc85-b57c-447f-ab4d-1294dcd0f61f","Type":"ContainerStarted","Data":"76934a632b4232e8d3eb27d19b267928612c866b1f8386bc853a43ba709d9b27"} Oct 05 09:17:45 crc kubenswrapper[4935]: I1005 09:17:45.720539 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-bqz75" event={"ID":"9113dc85-b57c-447f-ab4d-1294dcd0f61f","Type":"ContainerStarted","Data":"e7ab0117cd455e7a8e0826ab8055d5884dbdfe83a0332fad8c37d0b2d7916355"} Oct 05 09:17:45 crc kubenswrapper[4935]: I1005 09:17:45.756918 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-bqz75" podStartSLOduration=2.323681943 podStartE2EDuration="2.756879839s" podCreationTimestamp="2025-10-05 09:17:43 +0000 UTC" firstStartedPulling="2025-10-05 09:17:44.722301297 +0000 UTC m=+8698.604927757" lastFinishedPulling="2025-10-05 09:17:45.155499203 +0000 UTC m=+8699.038125653" observedRunningTime="2025-10-05 09:17:45.750937163 +0000 UTC m=+8699.633563623" watchObservedRunningTime="2025-10-05 09:17:45.756879839 +0000 UTC m=+8699.639506299" Oct 05 09:17:46 crc kubenswrapper[4935]: I1005 09:17:46.731989 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mm6wt" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerName="registry-server" containerID="cri-o://0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3" gracePeriod=2 Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.290144 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.445758 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-catalog-content\") pod \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.445993 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wp9v\" (UniqueName: \"kubernetes.io/projected/6b6e78d1-72d5-44de-9d15-827b2303e6a0-kube-api-access-9wp9v\") pod \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.446190 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-utilities\") pod \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\" (UID: \"6b6e78d1-72d5-44de-9d15-827b2303e6a0\") " Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.447593 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-utilities" (OuterVolumeSpecName: "utilities") pod "6b6e78d1-72d5-44de-9d15-827b2303e6a0" (UID: "6b6e78d1-72d5-44de-9d15-827b2303e6a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.461154 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b6e78d1-72d5-44de-9d15-827b2303e6a0-kube-api-access-9wp9v" (OuterVolumeSpecName: "kube-api-access-9wp9v") pod "6b6e78d1-72d5-44de-9d15-827b2303e6a0" (UID: "6b6e78d1-72d5-44de-9d15-827b2303e6a0"). InnerVolumeSpecName "kube-api-access-9wp9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.498828 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b6e78d1-72d5-44de-9d15-827b2303e6a0" (UID: "6b6e78d1-72d5-44de-9d15-827b2303e6a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.548965 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.549003 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b6e78d1-72d5-44de-9d15-827b2303e6a0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.549017 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wp9v\" (UniqueName: \"kubernetes.io/projected/6b6e78d1-72d5-44de-9d15-827b2303e6a0-kube-api-access-9wp9v\") on node \"crc\" DevicePath \"\"" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.762755 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerID="0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3" exitCode=0 Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.762802 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm6wt" event={"ID":"6b6e78d1-72d5-44de-9d15-827b2303e6a0","Type":"ContainerDied","Data":"0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3"} Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.762833 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm6wt" event={"ID":"6b6e78d1-72d5-44de-9d15-827b2303e6a0","Type":"ContainerDied","Data":"6d75e078e0d0580c40c4bf66f7b46fdbd5dc682f14c9fa7e536772b03dc9a3cc"} Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.762853 4935 scope.go:117] "RemoveContainer" containerID="0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.762938 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mm6wt" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.795207 4935 scope.go:117] "RemoveContainer" containerID="0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.820684 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mm6wt"] Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.832709 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mm6wt"] Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.842775 4935 scope.go:117] "RemoveContainer" containerID="cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.879955 4935 scope.go:117] "RemoveContainer" containerID="0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3" Oct 05 09:17:47 crc kubenswrapper[4935]: E1005 09:17:47.885399 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3\": container with ID starting with 0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3 not found: ID does not exist" containerID="0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.885469 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3"} err="failed to get container status \"0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3\": rpc error: code = NotFound desc = could not find container \"0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3\": container with ID starting with 0a9ad5c131b1d23cca2624b2049912ecb440e737d7af35f48d4c92a03774f4d3 not found: ID does not exist" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.885500 4935 scope.go:117] "RemoveContainer" containerID="0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7" Oct 05 09:17:47 crc kubenswrapper[4935]: E1005 09:17:47.885852 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7\": container with ID starting with 0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7 not found: ID does not exist" containerID="0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.885915 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7"} err="failed to get container status \"0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7\": rpc error: code = NotFound desc = could not find container \"0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7\": container with ID starting with 0eaf3eaadcbe01b27248c64ef9b67a9bd68b0540a3978c9bf0fef00a5f8b14e7 not found: ID does not exist" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.885937 4935 scope.go:117] "RemoveContainer" containerID="cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2" Oct 05 09:17:47 crc kubenswrapper[4935]: E1005 09:17:47.886208 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2\": container with ID starting with cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2 not found: ID does not exist" containerID="cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2" Oct 05 09:17:47 crc kubenswrapper[4935]: I1005 09:17:47.886242 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2"} err="failed to get container status \"cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2\": rpc error: code = NotFound desc = could not find container \"cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2\": container with ID starting with cfda526a91eab46e91f8ee901d4dccac79714fd34a3bab9d4472ec1234ed40a2 not found: ID does not exist" Oct 05 09:17:48 crc kubenswrapper[4935]: I1005 09:17:48.798049 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" path="/var/lib/kubelet/pods/6b6e78d1-72d5-44de-9d15-827b2303e6a0/volumes" Oct 05 09:18:23 crc kubenswrapper[4935]: I1005 09:18:23.182755 4935 generic.go:334] "Generic (PLEG): container finished" podID="cdbef832-881e-4a2a-8735-59f801251dd5" containerID="7c29d2c0a0459c3a95c3c3eb5c2c6b31bccfc0220a7f85e303a4628c21b84d91" exitCode=0 Oct 05 09:18:23 crc kubenswrapper[4935]: I1005 09:18:23.182822 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-xkr98" event={"ID":"cdbef832-881e-4a2a-8735-59f801251dd5","Type":"ContainerDied","Data":"7c29d2c0a0459c3a95c3c3eb5c2c6b31bccfc0220a7f85e303a4628c21b84d91"} Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.685830 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.716497 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s268\" (UniqueName: \"kubernetes.io/projected/cdbef832-881e-4a2a-8735-59f801251dd5-kube-api-access-6s268\") pod \"cdbef832-881e-4a2a-8735-59f801251dd5\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.716566 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-ssh-key\") pod \"cdbef832-881e-4a2a-8735-59f801251dd5\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.716722 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-inventory\") pod \"cdbef832-881e-4a2a-8735-59f801251dd5\" (UID: \"cdbef832-881e-4a2a-8735-59f801251dd5\") " Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.723615 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdbef832-881e-4a2a-8735-59f801251dd5-kube-api-access-6s268" (OuterVolumeSpecName: "kube-api-access-6s268") pod "cdbef832-881e-4a2a-8735-59f801251dd5" (UID: "cdbef832-881e-4a2a-8735-59f801251dd5"). InnerVolumeSpecName "kube-api-access-6s268". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.765813 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-inventory" (OuterVolumeSpecName: "inventory") pod "cdbef832-881e-4a2a-8735-59f801251dd5" (UID: "cdbef832-881e-4a2a-8735-59f801251dd5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.771945 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cdbef832-881e-4a2a-8735-59f801251dd5" (UID: "cdbef832-881e-4a2a-8735-59f801251dd5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.821772 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.821828 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s268\" (UniqueName: \"kubernetes.io/projected/cdbef832-881e-4a2a-8735-59f801251dd5-kube-api-access-6s268\") on node \"crc\" DevicePath \"\"" Oct 05 09:18:24 crc kubenswrapper[4935]: I1005 09:18:24.821858 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdbef832-881e-4a2a-8735-59f801251dd5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.203406 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-xkr98" event={"ID":"cdbef832-881e-4a2a-8735-59f801251dd5","Type":"ContainerDied","Data":"1745353a8ca0068f7e1bc8367d92501d9910069c0a71121da5a7b3c9235bf10a"} Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.203459 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1745353a8ca0068f7e1bc8367d92501d9910069c0a71121da5a7b3c9235bf10a" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.203557 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-xkr98" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.312734 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-networker-4cgvb"] Oct 05 09:18:25 crc kubenswrapper[4935]: E1005 09:18:25.313560 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerName="extract-utilities" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.313582 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerName="extract-utilities" Oct 05 09:18:25 crc kubenswrapper[4935]: E1005 09:18:25.313611 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerName="extract-content" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.313619 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerName="extract-content" Oct 05 09:18:25 crc kubenswrapper[4935]: E1005 09:18:25.313646 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdbef832-881e-4a2a-8735-59f801251dd5" containerName="install-os-openstack-openstack-networker" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.313656 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdbef832-881e-4a2a-8735-59f801251dd5" containerName="install-os-openstack-openstack-networker" Oct 05 09:18:25 crc kubenswrapper[4935]: E1005 09:18:25.314217 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerName="registry-server" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.314241 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerName="registry-server" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.314519 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b6e78d1-72d5-44de-9d15-827b2303e6a0" containerName="registry-server" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.314560 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdbef832-881e-4a2a-8735-59f801251dd5" containerName="install-os-openstack-openstack-networker" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.315531 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.317862 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.318507 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.333680 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-ssh-key\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.333875 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-inventory\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.333924 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdvfj\" (UniqueName: \"kubernetes.io/projected/abe126de-f191-4efd-b1cc-dfaa5937bc77-kube-api-access-cdvfj\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.340960 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-4cgvb"] Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.435972 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-ssh-key\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.436172 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-inventory\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.436222 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdvfj\" (UniqueName: \"kubernetes.io/projected/abe126de-f191-4efd-b1cc-dfaa5937bc77-kube-api-access-cdvfj\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.441438 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-inventory\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.443117 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-ssh-key\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.451163 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdvfj\" (UniqueName: \"kubernetes.io/projected/abe126de-f191-4efd-b1cc-dfaa5937bc77-kube-api-access-cdvfj\") pod \"configure-os-openstack-openstack-networker-4cgvb\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:25 crc kubenswrapper[4935]: I1005 09:18:25.637850 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:18:26 crc kubenswrapper[4935]: I1005 09:18:26.243563 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-4cgvb"] Oct 05 09:18:27 crc kubenswrapper[4935]: I1005 09:18:27.234943 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" event={"ID":"abe126de-f191-4efd-b1cc-dfaa5937bc77","Type":"ContainerStarted","Data":"6c01c8dafebaa831fa9dcad10c69f421ad4e2c52b9b029b921a1c0e063c34e14"} Oct 05 09:18:27 crc kubenswrapper[4935]: I1005 09:18:27.235767 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" event={"ID":"abe126de-f191-4efd-b1cc-dfaa5937bc77","Type":"ContainerStarted","Data":"60d96327e4e07a24f4f7c57247ac80029c70aad2d14dc187eef82d4f0342b273"} Oct 05 09:18:27 crc kubenswrapper[4935]: I1005 09:18:27.274868 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" podStartSLOduration=1.794370574 podStartE2EDuration="2.274842562s" podCreationTimestamp="2025-10-05 09:18:25 +0000 UTC" firstStartedPulling="2025-10-05 09:18:26.267532278 +0000 UTC m=+8740.150158738" lastFinishedPulling="2025-10-05 09:18:26.748004226 +0000 UTC m=+8740.630630726" observedRunningTime="2025-10-05 09:18:27.259277713 +0000 UTC m=+8741.141904213" watchObservedRunningTime="2025-10-05 09:18:27.274842562 +0000 UTC m=+8741.157469062" Oct 05 09:18:33 crc kubenswrapper[4935]: I1005 09:18:33.303868 4935 generic.go:334] "Generic (PLEG): container finished" podID="9113dc85-b57c-447f-ab4d-1294dcd0f61f" containerID="76934a632b4232e8d3eb27d19b267928612c866b1f8386bc853a43ba709d9b27" exitCode=0 Oct 05 09:18:33 crc kubenswrapper[4935]: I1005 09:18:33.304024 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-bqz75" event={"ID":"9113dc85-b57c-447f-ab4d-1294dcd0f61f","Type":"ContainerDied","Data":"76934a632b4232e8d3eb27d19b267928612c866b1f8386bc853a43ba709d9b27"} Oct 05 09:18:34 crc kubenswrapper[4935]: I1005 09:18:34.810690 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:18:34 crc kubenswrapper[4935]: I1005 09:18:34.952855 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s92p\" (UniqueName: \"kubernetes.io/projected/9113dc85-b57c-447f-ab4d-1294dcd0f61f-kube-api-access-6s92p\") pod \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " Oct 05 09:18:34 crc kubenswrapper[4935]: I1005 09:18:34.953007 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ceph\") pod \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " Oct 05 09:18:34 crc kubenswrapper[4935]: I1005 09:18:34.953083 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ssh-key\") pod \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " Oct 05 09:18:34 crc kubenswrapper[4935]: I1005 09:18:34.953131 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-inventory\") pod \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\" (UID: \"9113dc85-b57c-447f-ab4d-1294dcd0f61f\") " Oct 05 09:18:34 crc kubenswrapper[4935]: I1005 09:18:34.958881 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ceph" (OuterVolumeSpecName: "ceph") pod "9113dc85-b57c-447f-ab4d-1294dcd0f61f" (UID: "9113dc85-b57c-447f-ab4d-1294dcd0f61f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:18:34 crc kubenswrapper[4935]: I1005 09:18:34.959143 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9113dc85-b57c-447f-ab4d-1294dcd0f61f-kube-api-access-6s92p" (OuterVolumeSpecName: "kube-api-access-6s92p") pod "9113dc85-b57c-447f-ab4d-1294dcd0f61f" (UID: "9113dc85-b57c-447f-ab4d-1294dcd0f61f"). InnerVolumeSpecName "kube-api-access-6s92p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.002190 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9113dc85-b57c-447f-ab4d-1294dcd0f61f" (UID: "9113dc85-b57c-447f-ab4d-1294dcd0f61f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.002988 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-inventory" (OuterVolumeSpecName: "inventory") pod "9113dc85-b57c-447f-ab4d-1294dcd0f61f" (UID: "9113dc85-b57c-447f-ab4d-1294dcd0f61f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.055366 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.055409 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.055424 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113dc85-b57c-447f-ab4d-1294dcd0f61f-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.055438 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s92p\" (UniqueName: \"kubernetes.io/projected/9113dc85-b57c-447f-ab4d-1294dcd0f61f-kube-api-access-6s92p\") on node \"crc\" DevicePath \"\"" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.331532 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-bqz75" event={"ID":"9113dc85-b57c-447f-ab4d-1294dcd0f61f","Type":"ContainerDied","Data":"e7ab0117cd455e7a8e0826ab8055d5884dbdfe83a0332fad8c37d0b2d7916355"} Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.331576 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7ab0117cd455e7a8e0826ab8055d5884dbdfe83a0332fad8c37d0b2d7916355" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.331621 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-bqz75" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.422017 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-cn2bk"] Oct 05 09:18:35 crc kubenswrapper[4935]: E1005 09:18:35.422529 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9113dc85-b57c-447f-ab4d-1294dcd0f61f" containerName="install-os-openstack-openstack-cell1" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.422551 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9113dc85-b57c-447f-ab4d-1294dcd0f61f" containerName="install-os-openstack-openstack-cell1" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.422821 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9113dc85-b57c-447f-ab4d-1294dcd0f61f" containerName="install-os-openstack-openstack-cell1" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.426448 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.428812 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.430076 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.440720 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-cn2bk"] Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.566612 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ceph\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.566688 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ssh-key\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.566786 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-inventory\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.566836 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdlcn\" (UniqueName: \"kubernetes.io/projected/141a4fc2-2a6a-4464-bd8a-9ad826abba58-kube-api-access-wdlcn\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.669266 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-inventory\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.669323 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdlcn\" (UniqueName: \"kubernetes.io/projected/141a4fc2-2a6a-4464-bd8a-9ad826abba58-kube-api-access-wdlcn\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.669421 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ceph\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.669465 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ssh-key\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.673967 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ssh-key\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.674253 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ceph\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.676651 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-inventory\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.687743 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdlcn\" (UniqueName: \"kubernetes.io/projected/141a4fc2-2a6a-4464-bd8a-9ad826abba58-kube-api-access-wdlcn\") pod \"configure-os-openstack-openstack-cell1-cn2bk\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:35 crc kubenswrapper[4935]: I1005 09:18:35.746483 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:18:36 crc kubenswrapper[4935]: I1005 09:18:36.333202 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-cn2bk"] Oct 05 09:18:37 crc kubenswrapper[4935]: I1005 09:18:37.353147 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" event={"ID":"141a4fc2-2a6a-4464-bd8a-9ad826abba58","Type":"ContainerStarted","Data":"117acb0587a38f47d35ee368e321480bba561189edfa67d006d5b4d5725face8"} Oct 05 09:18:37 crc kubenswrapper[4935]: I1005 09:18:37.353915 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" event={"ID":"141a4fc2-2a6a-4464-bd8a-9ad826abba58","Type":"ContainerStarted","Data":"8f68a820a5fd63820b1d78fabf61d58f8d7b11e377c91e0cb50fe31b1cd102c0"} Oct 05 09:18:37 crc kubenswrapper[4935]: I1005 09:18:37.385158 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" podStartSLOduration=1.881576533 podStartE2EDuration="2.385127128s" podCreationTimestamp="2025-10-05 09:18:35 +0000 UTC" firstStartedPulling="2025-10-05 09:18:36.337256417 +0000 UTC m=+8750.219882877" lastFinishedPulling="2025-10-05 09:18:36.840807012 +0000 UTC m=+8750.723433472" observedRunningTime="2025-10-05 09:18:37.371062498 +0000 UTC m=+8751.253688978" watchObservedRunningTime="2025-10-05 09:18:37.385127128 +0000 UTC m=+8751.267753588" Oct 05 09:19:21 crc kubenswrapper[4935]: I1005 09:19:21.831893 4935 generic.go:334] "Generic (PLEG): container finished" podID="abe126de-f191-4efd-b1cc-dfaa5937bc77" containerID="6c01c8dafebaa831fa9dcad10c69f421ad4e2c52b9b029b921a1c0e063c34e14" exitCode=2 Oct 05 09:19:21 crc kubenswrapper[4935]: I1005 09:19:21.832016 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" event={"ID":"abe126de-f191-4efd-b1cc-dfaa5937bc77","Type":"ContainerDied","Data":"6c01c8dafebaa831fa9dcad10c69f421ad4e2c52b9b029b921a1c0e063c34e14"} Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.343620 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.480300 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdvfj\" (UniqueName: \"kubernetes.io/projected/abe126de-f191-4efd-b1cc-dfaa5937bc77-kube-api-access-cdvfj\") pod \"abe126de-f191-4efd-b1cc-dfaa5937bc77\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.480569 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-ssh-key\") pod \"abe126de-f191-4efd-b1cc-dfaa5937bc77\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.480643 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-inventory\") pod \"abe126de-f191-4efd-b1cc-dfaa5937bc77\" (UID: \"abe126de-f191-4efd-b1cc-dfaa5937bc77\") " Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.490621 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe126de-f191-4efd-b1cc-dfaa5937bc77-kube-api-access-cdvfj" (OuterVolumeSpecName: "kube-api-access-cdvfj") pod "abe126de-f191-4efd-b1cc-dfaa5937bc77" (UID: "abe126de-f191-4efd-b1cc-dfaa5937bc77"). InnerVolumeSpecName "kube-api-access-cdvfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.513254 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-inventory" (OuterVolumeSpecName: "inventory") pod "abe126de-f191-4efd-b1cc-dfaa5937bc77" (UID: "abe126de-f191-4efd-b1cc-dfaa5937bc77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.521490 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "abe126de-f191-4efd-b1cc-dfaa5937bc77" (UID: "abe126de-f191-4efd-b1cc-dfaa5937bc77"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.585030 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdvfj\" (UniqueName: \"kubernetes.io/projected/abe126de-f191-4efd-b1cc-dfaa5937bc77-kube-api-access-cdvfj\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.585491 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.585568 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe126de-f191-4efd-b1cc-dfaa5937bc77-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.869809 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" event={"ID":"abe126de-f191-4efd-b1cc-dfaa5937bc77","Type":"ContainerDied","Data":"60d96327e4e07a24f4f7c57247ac80029c70aad2d14dc187eef82d4f0342b273"} Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.870399 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60d96327e4e07a24f4f7c57247ac80029c70aad2d14dc187eef82d4f0342b273" Oct 05 09:19:23 crc kubenswrapper[4935]: I1005 09:19:23.870091 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-4cgvb" Oct 05 09:19:30 crc kubenswrapper[4935]: I1005 09:19:30.945663 4935 generic.go:334] "Generic (PLEG): container finished" podID="141a4fc2-2a6a-4464-bd8a-9ad826abba58" containerID="117acb0587a38f47d35ee368e321480bba561189edfa67d006d5b4d5725face8" exitCode=2 Oct 05 09:19:30 crc kubenswrapper[4935]: I1005 09:19:30.945923 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" event={"ID":"141a4fc2-2a6a-4464-bd8a-9ad826abba58","Type":"ContainerDied","Data":"117acb0587a38f47d35ee368e321480bba561189edfa67d006d5b4d5725face8"} Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.026651 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-networker-7nmjq"] Oct 05 09:19:31 crc kubenswrapper[4935]: E1005 09:19:31.027108 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe126de-f191-4efd-b1cc-dfaa5937bc77" containerName="configure-os-openstack-openstack-networker" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.027125 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe126de-f191-4efd-b1cc-dfaa5937bc77" containerName="configure-os-openstack-openstack-networker" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.027362 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="abe126de-f191-4efd-b1cc-dfaa5937bc77" containerName="configure-os-openstack-openstack-networker" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.028253 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.030008 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.031846 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.035681 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-7nmjq"] Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.142725 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-ssh-key\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.143209 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g7rl\" (UniqueName: \"kubernetes.io/projected/f8b0de95-9148-4f8d-aab4-997b20290690-kube-api-access-9g7rl\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.143435 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-inventory\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.246616 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g7rl\" (UniqueName: \"kubernetes.io/projected/f8b0de95-9148-4f8d-aab4-997b20290690-kube-api-access-9g7rl\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.246727 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-inventory\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.246952 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-ssh-key\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.562834 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-ssh-key\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.563450 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-inventory\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.563526 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g7rl\" (UniqueName: \"kubernetes.io/projected/f8b0de95-9148-4f8d-aab4-997b20290690-kube-api-access-9g7rl\") pod \"configure-os-openstack-openstack-networker-7nmjq\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:31 crc kubenswrapper[4935]: I1005 09:19:31.680507 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.284702 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-7nmjq"] Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.452840 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.586858 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-inventory\") pod \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.587131 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ceph\") pod \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.587242 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdlcn\" (UniqueName: \"kubernetes.io/projected/141a4fc2-2a6a-4464-bd8a-9ad826abba58-kube-api-access-wdlcn\") pod \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.587285 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ssh-key\") pod \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\" (UID: \"141a4fc2-2a6a-4464-bd8a-9ad826abba58\") " Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.594672 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/141a4fc2-2a6a-4464-bd8a-9ad826abba58-kube-api-access-wdlcn" (OuterVolumeSpecName: "kube-api-access-wdlcn") pod "141a4fc2-2a6a-4464-bd8a-9ad826abba58" (UID: "141a4fc2-2a6a-4464-bd8a-9ad826abba58"). InnerVolumeSpecName "kube-api-access-wdlcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.607487 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ceph" (OuterVolumeSpecName: "ceph") pod "141a4fc2-2a6a-4464-bd8a-9ad826abba58" (UID: "141a4fc2-2a6a-4464-bd8a-9ad826abba58"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.623695 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-inventory" (OuterVolumeSpecName: "inventory") pod "141a4fc2-2a6a-4464-bd8a-9ad826abba58" (UID: "141a4fc2-2a6a-4464-bd8a-9ad826abba58"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.632933 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "141a4fc2-2a6a-4464-bd8a-9ad826abba58" (UID: "141a4fc2-2a6a-4464-bd8a-9ad826abba58"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.690367 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.690393 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdlcn\" (UniqueName: \"kubernetes.io/projected/141a4fc2-2a6a-4464-bd8a-9ad826abba58-kube-api-access-wdlcn\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.690406 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.690415 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/141a4fc2-2a6a-4464-bd8a-9ad826abba58-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.971501 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" event={"ID":"141a4fc2-2a6a-4464-bd8a-9ad826abba58","Type":"ContainerDied","Data":"8f68a820a5fd63820b1d78fabf61d58f8d7b11e377c91e0cb50fe31b1cd102c0"} Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.971542 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f68a820a5fd63820b1d78fabf61d58f8d7b11e377c91e0cb50fe31b1cd102c0" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.971668 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-cn2bk" Oct 05 09:19:32 crc kubenswrapper[4935]: I1005 09:19:32.973578 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" event={"ID":"f8b0de95-9148-4f8d-aab4-997b20290690","Type":"ContainerStarted","Data":"70afbc25d9a6f9e402c3fe7d17327123cc8021f4b930823da3b30c2129cfe87a"} Oct 05 09:19:33 crc kubenswrapper[4935]: I1005 09:19:33.986554 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" event={"ID":"f8b0de95-9148-4f8d-aab4-997b20290690","Type":"ContainerStarted","Data":"42a5f209c00c2eb3f4ac97e8c17c7a018f896225981b7f830c22b644711ccf5d"} Oct 05 09:19:34 crc kubenswrapper[4935]: I1005 09:19:34.017965 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" podStartSLOduration=2.43165549 podStartE2EDuration="3.01794618s" podCreationTimestamp="2025-10-05 09:19:31 +0000 UTC" firstStartedPulling="2025-10-05 09:19:32.305017519 +0000 UTC m=+8806.187643969" lastFinishedPulling="2025-10-05 09:19:32.891308199 +0000 UTC m=+8806.773934659" observedRunningTime="2025-10-05 09:19:34.01225573 +0000 UTC m=+8807.894882200" watchObservedRunningTime="2025-10-05 09:19:34.01794618 +0000 UTC m=+8807.900572640" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.024145 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-4cfll"] Oct 05 09:19:40 crc kubenswrapper[4935]: E1005 09:19:40.025117 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="141a4fc2-2a6a-4464-bd8a-9ad826abba58" containerName="configure-os-openstack-openstack-cell1" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.025132 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="141a4fc2-2a6a-4464-bd8a-9ad826abba58" containerName="configure-os-openstack-openstack-cell1" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.025323 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="141a4fc2-2a6a-4464-bd8a-9ad826abba58" containerName="configure-os-openstack-openstack-cell1" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.026065 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.033057 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.033280 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.039958 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-4cfll"] Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.142787 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94pgr\" (UniqueName: \"kubernetes.io/projected/697225c9-e0f0-46b7-9fc1-6456384033bb-kube-api-access-94pgr\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.142877 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ceph\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.142969 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-inventory\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.143094 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ssh-key\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.245917 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94pgr\" (UniqueName: \"kubernetes.io/projected/697225c9-e0f0-46b7-9fc1-6456384033bb-kube-api-access-94pgr\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.246401 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ceph\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.246594 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-inventory\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.246818 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ssh-key\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.256466 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ssh-key\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.256650 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-inventory\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.257273 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ceph\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.275704 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94pgr\" (UniqueName: \"kubernetes.io/projected/697225c9-e0f0-46b7-9fc1-6456384033bb-kube-api-access-94pgr\") pod \"configure-os-openstack-openstack-cell1-4cfll\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.364353 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:19:40 crc kubenswrapper[4935]: I1005 09:19:40.738937 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-4cfll"] Oct 05 09:19:42 crc kubenswrapper[4935]: I1005 09:19:42.066711 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" event={"ID":"697225c9-e0f0-46b7-9fc1-6456384033bb","Type":"ContainerStarted","Data":"d6f770620ded27c9b725f48e619c50cd715150f2c8a117e56276b7d5fce07a93"} Oct 05 09:19:42 crc kubenswrapper[4935]: I1005 09:19:42.067263 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" event={"ID":"697225c9-e0f0-46b7-9fc1-6456384033bb","Type":"ContainerStarted","Data":"48155bda8b73176dcc10b544b97fcaf3bb9a0af6a5022b1e85e2d490280ce565"} Oct 05 09:19:42 crc kubenswrapper[4935]: I1005 09:19:42.082746 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" podStartSLOduration=1.634880973 podStartE2EDuration="2.082728404s" podCreationTimestamp="2025-10-05 09:19:40 +0000 UTC" firstStartedPulling="2025-10-05 09:19:41.07013522 +0000 UTC m=+8814.952761680" lastFinishedPulling="2025-10-05 09:19:41.517982651 +0000 UTC m=+8815.400609111" observedRunningTime="2025-10-05 09:19:42.079508059 +0000 UTC m=+8815.962134519" watchObservedRunningTime="2025-10-05 09:19:42.082728404 +0000 UTC m=+8815.965354864" Oct 05 09:19:44 crc kubenswrapper[4935]: I1005 09:19:44.290064 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:19:44 crc kubenswrapper[4935]: I1005 09:19:44.291232 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.073216 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w5kzc"] Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.076443 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.098207 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5kzc"] Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.179907 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-utilities\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.179998 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-catalog-content\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.180117 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l442j\" (UniqueName: \"kubernetes.io/projected/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-kube-api-access-l442j\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.282154 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-utilities\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.282231 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-catalog-content\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.282309 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l442j\" (UniqueName: \"kubernetes.io/projected/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-kube-api-access-l442j\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.282701 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-utilities\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.282762 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-catalog-content\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.304707 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l442j\" (UniqueName: \"kubernetes.io/projected/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-kube-api-access-l442j\") pod \"redhat-marketplace-w5kzc\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.396223 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:13 crc kubenswrapper[4935]: I1005 09:20:13.891734 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5kzc"] Oct 05 09:20:14 crc kubenswrapper[4935]: I1005 09:20:14.289923 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:20:14 crc kubenswrapper[4935]: I1005 09:20:14.291187 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:20:14 crc kubenswrapper[4935]: I1005 09:20:14.442318 4935 generic.go:334] "Generic (PLEG): container finished" podID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerID="560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264" exitCode=0 Oct 05 09:20:14 crc kubenswrapper[4935]: I1005 09:20:14.442370 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5kzc" event={"ID":"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9","Type":"ContainerDied","Data":"560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264"} Oct 05 09:20:14 crc kubenswrapper[4935]: I1005 09:20:14.442411 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5kzc" event={"ID":"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9","Type":"ContainerStarted","Data":"dd03b02460575bdb6f354730532d238ff0e160bf0d54674b1f2e64df73bfa533"} Oct 05 09:20:14 crc kubenswrapper[4935]: I1005 09:20:14.444775 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:20:15 crc kubenswrapper[4935]: I1005 09:20:15.453222 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5kzc" event={"ID":"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9","Type":"ContainerStarted","Data":"ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e"} Oct 05 09:20:16 crc kubenswrapper[4935]: I1005 09:20:16.463053 4935 generic.go:334] "Generic (PLEG): container finished" podID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerID="ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e" exitCode=0 Oct 05 09:20:16 crc kubenswrapper[4935]: I1005 09:20:16.463313 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5kzc" event={"ID":"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9","Type":"ContainerDied","Data":"ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e"} Oct 05 09:20:17 crc kubenswrapper[4935]: I1005 09:20:17.473656 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5kzc" event={"ID":"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9","Type":"ContainerStarted","Data":"90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb"} Oct 05 09:20:17 crc kubenswrapper[4935]: I1005 09:20:17.493143 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w5kzc" podStartSLOduration=2.031952387 podStartE2EDuration="4.493123644s" podCreationTimestamp="2025-10-05 09:20:13 +0000 UTC" firstStartedPulling="2025-10-05 09:20:14.444568619 +0000 UTC m=+8848.327195069" lastFinishedPulling="2025-10-05 09:20:16.905739866 +0000 UTC m=+8850.788366326" observedRunningTime="2025-10-05 09:20:17.487175238 +0000 UTC m=+8851.369801698" watchObservedRunningTime="2025-10-05 09:20:17.493123644 +0000 UTC m=+8851.375750104" Oct 05 09:20:23 crc kubenswrapper[4935]: I1005 09:20:23.397407 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:23 crc kubenswrapper[4935]: I1005 09:20:23.398051 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:23 crc kubenswrapper[4935]: I1005 09:20:23.465577 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:23 crc kubenswrapper[4935]: I1005 09:20:23.526268 4935 generic.go:334] "Generic (PLEG): container finished" podID="f8b0de95-9148-4f8d-aab4-997b20290690" containerID="42a5f209c00c2eb3f4ac97e8c17c7a018f896225981b7f830c22b644711ccf5d" exitCode=0 Oct 05 09:20:23 crc kubenswrapper[4935]: I1005 09:20:23.526330 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" event={"ID":"f8b0de95-9148-4f8d-aab4-997b20290690","Type":"ContainerDied","Data":"42a5f209c00c2eb3f4ac97e8c17c7a018f896225981b7f830c22b644711ccf5d"} Oct 05 09:20:23 crc kubenswrapper[4935]: I1005 09:20:23.575207 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:23 crc kubenswrapper[4935]: I1005 09:20:23.709878 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5kzc"] Oct 05 09:20:24 crc kubenswrapper[4935]: I1005 09:20:24.945402 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.034464 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-inventory\") pod \"f8b0de95-9148-4f8d-aab4-997b20290690\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.034612 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-ssh-key\") pod \"f8b0de95-9148-4f8d-aab4-997b20290690\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.034650 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g7rl\" (UniqueName: \"kubernetes.io/projected/f8b0de95-9148-4f8d-aab4-997b20290690-kube-api-access-9g7rl\") pod \"f8b0de95-9148-4f8d-aab4-997b20290690\" (UID: \"f8b0de95-9148-4f8d-aab4-997b20290690\") " Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.040123 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8b0de95-9148-4f8d-aab4-997b20290690-kube-api-access-9g7rl" (OuterVolumeSpecName: "kube-api-access-9g7rl") pod "f8b0de95-9148-4f8d-aab4-997b20290690" (UID: "f8b0de95-9148-4f8d-aab4-997b20290690"). InnerVolumeSpecName "kube-api-access-9g7rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.065916 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-inventory" (OuterVolumeSpecName: "inventory") pod "f8b0de95-9148-4f8d-aab4-997b20290690" (UID: "f8b0de95-9148-4f8d-aab4-997b20290690"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.074188 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f8b0de95-9148-4f8d-aab4-997b20290690" (UID: "f8b0de95-9148-4f8d-aab4-997b20290690"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.137332 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.137370 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8b0de95-9148-4f8d-aab4-997b20290690-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.137387 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g7rl\" (UniqueName: \"kubernetes.io/projected/f8b0de95-9148-4f8d-aab4-997b20290690-kube-api-access-9g7rl\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.543648 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.543680 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-7nmjq" event={"ID":"f8b0de95-9148-4f8d-aab4-997b20290690","Type":"ContainerDied","Data":"70afbc25d9a6f9e402c3fe7d17327123cc8021f4b930823da3b30c2129cfe87a"} Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.543743 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70afbc25d9a6f9e402c3fe7d17327123cc8021f4b930823da3b30c2129cfe87a" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.543767 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w5kzc" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerName="registry-server" containerID="cri-o://90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb" gracePeriod=2 Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.644455 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-networker-f2sdw"] Oct 05 09:20:25 crc kubenswrapper[4935]: E1005 09:20:25.645169 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b0de95-9148-4f8d-aab4-997b20290690" containerName="configure-os-openstack-openstack-networker" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.645188 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b0de95-9148-4f8d-aab4-997b20290690" containerName="configure-os-openstack-openstack-networker" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.645386 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8b0de95-9148-4f8d-aab4-997b20290690" containerName="configure-os-openstack-openstack-networker" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.646171 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.649239 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.650625 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.664092 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-f2sdw"] Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.751662 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-ssh-key\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.751751 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-inventory\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.751942 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6qzq\" (UniqueName: \"kubernetes.io/projected/43c59bed-43ce-4cef-9d35-45e7496c92f3-kube-api-access-t6qzq\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.863396 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-ssh-key\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.863521 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-inventory\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.863796 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6qzq\" (UniqueName: \"kubernetes.io/projected/43c59bed-43ce-4cef-9d35-45e7496c92f3-kube-api-access-t6qzq\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.868504 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-inventory\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.869197 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-ssh-key\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.893178 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6qzq\" (UniqueName: \"kubernetes.io/projected/43c59bed-43ce-4cef-9d35-45e7496c92f3-kube-api-access-t6qzq\") pod \"run-os-openstack-openstack-networker-f2sdw\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:25 crc kubenswrapper[4935]: I1005 09:20:25.965840 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.012773 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.067078 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-utilities\") pod \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.067481 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-catalog-content\") pod \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.067610 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l442j\" (UniqueName: \"kubernetes.io/projected/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-kube-api-access-l442j\") pod \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\" (UID: \"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9\") " Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.069094 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-utilities" (OuterVolumeSpecName: "utilities") pod "6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" (UID: "6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.069535 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.074788 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-kube-api-access-l442j" (OuterVolumeSpecName: "kube-api-access-l442j") pod "6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" (UID: "6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9"). InnerVolumeSpecName "kube-api-access-l442j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.085650 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" (UID: "6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.171099 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.171138 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l442j\" (UniqueName: \"kubernetes.io/projected/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9-kube-api-access-l442j\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.538518 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-f2sdw"] Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.559419 4935 generic.go:334] "Generic (PLEG): container finished" podID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerID="90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb" exitCode=0 Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.559481 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5kzc" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.559508 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5kzc" event={"ID":"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9","Type":"ContainerDied","Data":"90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb"} Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.559542 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5kzc" event={"ID":"6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9","Type":"ContainerDied","Data":"dd03b02460575bdb6f354730532d238ff0e160bf0d54674b1f2e64df73bfa533"} Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.559563 4935 scope.go:117] "RemoveContainer" containerID="90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.562058 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-f2sdw" event={"ID":"43c59bed-43ce-4cef-9d35-45e7496c92f3","Type":"ContainerStarted","Data":"99071115e55e6ef7625bc71950dde5fb7cdef9b8036536a15f07c810221bace2"} Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.606231 4935 scope.go:117] "RemoveContainer" containerID="ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.612565 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5kzc"] Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.660329 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5kzc"] Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.720354 4935 scope.go:117] "RemoveContainer" containerID="560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.795672 4935 scope.go:117] "RemoveContainer" containerID="90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb" Oct 05 09:20:26 crc kubenswrapper[4935]: E1005 09:20:26.800048 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb\": container with ID starting with 90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb not found: ID does not exist" containerID="90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.800093 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb"} err="failed to get container status \"90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb\": rpc error: code = NotFound desc = could not find container \"90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb\": container with ID starting with 90617688dae3cf4d00c81e5a256f46a614805b4c070af7c28d907a76513722fb not found: ID does not exist" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.800118 4935 scope.go:117] "RemoveContainer" containerID="ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e" Oct 05 09:20:26 crc kubenswrapper[4935]: E1005 09:20:26.804352 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e\": container with ID starting with ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e not found: ID does not exist" containerID="ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.804393 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e"} err="failed to get container status \"ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e\": rpc error: code = NotFound desc = could not find container \"ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e\": container with ID starting with ff27718ba76c1f54745b2dea4a12e7c8989ace45c0a7f7191bc3c2efd48fa22e not found: ID does not exist" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.804416 4935 scope.go:117] "RemoveContainer" containerID="560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264" Oct 05 09:20:26 crc kubenswrapper[4935]: E1005 09:20:26.812370 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264\": container with ID starting with 560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264 not found: ID does not exist" containerID="560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.812413 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264"} err="failed to get container status \"560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264\": rpc error: code = NotFound desc = could not find container \"560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264\": container with ID starting with 560d76a2e2e8957b1e7989dcce4eb17602e54cb901366f270b5915147d60d264 not found: ID does not exist" Oct 05 09:20:26 crc kubenswrapper[4935]: I1005 09:20:26.824266 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" path="/var/lib/kubelet/pods/6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9/volumes" Oct 05 09:20:27 crc kubenswrapper[4935]: I1005 09:20:27.574477 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-f2sdw" event={"ID":"43c59bed-43ce-4cef-9d35-45e7496c92f3","Type":"ContainerStarted","Data":"1460fd0eae980d2b38f57f12269f65be81faa7b71c9a90b92dd85028e4a00843"} Oct 05 09:20:27 crc kubenswrapper[4935]: I1005 09:20:27.595941 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-networker-f2sdw" podStartSLOduration=2.135347397 podStartE2EDuration="2.595920312s" podCreationTimestamp="2025-10-05 09:20:25 +0000 UTC" firstStartedPulling="2025-10-05 09:20:26.542867475 +0000 UTC m=+8860.425493935" lastFinishedPulling="2025-10-05 09:20:27.00344039 +0000 UTC m=+8860.886066850" observedRunningTime="2025-10-05 09:20:27.588516158 +0000 UTC m=+8861.471142618" watchObservedRunningTime="2025-10-05 09:20:27.595920312 +0000 UTC m=+8861.478546772" Oct 05 09:20:28 crc kubenswrapper[4935]: I1005 09:20:28.594306 4935 generic.go:334] "Generic (PLEG): container finished" podID="697225c9-e0f0-46b7-9fc1-6456384033bb" containerID="d6f770620ded27c9b725f48e619c50cd715150f2c8a117e56276b7d5fce07a93" exitCode=0 Oct 05 09:20:28 crc kubenswrapper[4935]: I1005 09:20:28.594437 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" event={"ID":"697225c9-e0f0-46b7-9fc1-6456384033bb","Type":"ContainerDied","Data":"d6f770620ded27c9b725f48e619c50cd715150f2c8a117e56276b7d5fce07a93"} Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.809601 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.877489 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94pgr\" (UniqueName: \"kubernetes.io/projected/697225c9-e0f0-46b7-9fc1-6456384033bb-kube-api-access-94pgr\") pod \"697225c9-e0f0-46b7-9fc1-6456384033bb\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.878043 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ssh-key\") pod \"697225c9-e0f0-46b7-9fc1-6456384033bb\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.878403 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-inventory\") pod \"697225c9-e0f0-46b7-9fc1-6456384033bb\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.878458 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ceph\") pod \"697225c9-e0f0-46b7-9fc1-6456384033bb\" (UID: \"697225c9-e0f0-46b7-9fc1-6456384033bb\") " Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.883044 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ceph" (OuterVolumeSpecName: "ceph") pod "697225c9-e0f0-46b7-9fc1-6456384033bb" (UID: "697225c9-e0f0-46b7-9fc1-6456384033bb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.883335 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/697225c9-e0f0-46b7-9fc1-6456384033bb-kube-api-access-94pgr" (OuterVolumeSpecName: "kube-api-access-94pgr") pod "697225c9-e0f0-46b7-9fc1-6456384033bb" (UID: "697225c9-e0f0-46b7-9fc1-6456384033bb"). InnerVolumeSpecName "kube-api-access-94pgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.906101 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "697225c9-e0f0-46b7-9fc1-6456384033bb" (UID: "697225c9-e0f0-46b7-9fc1-6456384033bb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.911563 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-inventory" (OuterVolumeSpecName: "inventory") pod "697225c9-e0f0-46b7-9fc1-6456384033bb" (UID: "697225c9-e0f0-46b7-9fc1-6456384033bb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.981072 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.981106 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.981115 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/697225c9-e0f0-46b7-9fc1-6456384033bb-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:30 crc kubenswrapper[4935]: I1005 09:20:30.981124 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94pgr\" (UniqueName: \"kubernetes.io/projected/697225c9-e0f0-46b7-9fc1-6456384033bb-kube-api-access-94pgr\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.628196 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" event={"ID":"697225c9-e0f0-46b7-9fc1-6456384033bb","Type":"ContainerDied","Data":"48155bda8b73176dcc10b544b97fcaf3bb9a0af6a5022b1e85e2d490280ce565"} Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.628532 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48155bda8b73176dcc10b544b97fcaf3bb9a0af6a5022b1e85e2d490280ce565" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.628246 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-4cfll" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.940157 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-jn258"] Oct 05 09:20:31 crc kubenswrapper[4935]: E1005 09:20:31.940651 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerName="registry-server" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.940667 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerName="registry-server" Oct 05 09:20:31 crc kubenswrapper[4935]: E1005 09:20:31.940689 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerName="extract-content" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.940699 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerName="extract-content" Oct 05 09:20:31 crc kubenswrapper[4935]: E1005 09:20:31.940741 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerName="extract-utilities" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.940749 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerName="extract-utilities" Oct 05 09:20:31 crc kubenswrapper[4935]: E1005 09:20:31.940776 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="697225c9-e0f0-46b7-9fc1-6456384033bb" containerName="configure-os-openstack-openstack-cell1" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.940785 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="697225c9-e0f0-46b7-9fc1-6456384033bb" containerName="configure-os-openstack-openstack-cell1" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.941088 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="697225c9-e0f0-46b7-9fc1-6456384033bb" containerName="configure-os-openstack-openstack-cell1" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.941113 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dd0d26e-6464-4fbb-8ed5-0b9658ef02c9" containerName="registry-server" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.941943 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.944829 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.945126 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:20:31 crc kubenswrapper[4935]: I1005 09:20:31.947104 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-jn258"] Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.007298 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-1\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.007396 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnpzf\" (UniqueName: \"kubernetes.io/projected/06081368-97c9-41f4-9fed-75c593994b11-kube-api-access-nnpzf\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.007504 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ceph\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.007573 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.007629 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-0\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.007753 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.110046 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.110198 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-1\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.110240 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnpzf\" (UniqueName: \"kubernetes.io/projected/06081368-97c9-41f4-9fed-75c593994b11-kube-api-access-nnpzf\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.110306 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ceph\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.110359 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.110404 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-0\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.116003 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ceph\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.116670 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-0\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.116822 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.117314 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-1\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.122349 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.131239 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnpzf\" (UniqueName: \"kubernetes.io/projected/06081368-97c9-41f4-9fed-75c593994b11-kube-api-access-nnpzf\") pod \"ssh-known-hosts-openstack-jn258\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:32 crc kubenswrapper[4935]: I1005 09:20:32.267403 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:33 crc kubenswrapper[4935]: I1005 09:20:33.541864 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-jn258"] Oct 05 09:20:33 crc kubenswrapper[4935]: I1005 09:20:33.650330 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-jn258" event={"ID":"06081368-97c9-41f4-9fed-75c593994b11","Type":"ContainerStarted","Data":"0b91eb2445ae6e448d0c6287973a81adab6cbc2f0dcd54b6f54dc79bc6e17164"} Oct 05 09:20:34 crc kubenswrapper[4935]: I1005 09:20:34.661669 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-jn258" event={"ID":"06081368-97c9-41f4-9fed-75c593994b11","Type":"ContainerStarted","Data":"f7ea98626907a5b0aacae9bfd4e810283589bda010a88ece9a970d14d9551604"} Oct 05 09:20:34 crc kubenswrapper[4935]: I1005 09:20:34.689210 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-jn258" podStartSLOduration=3.229289424 podStartE2EDuration="3.689189132s" podCreationTimestamp="2025-10-05 09:20:31 +0000 UTC" firstStartedPulling="2025-10-05 09:20:33.55224971 +0000 UTC m=+8867.434876170" lastFinishedPulling="2025-10-05 09:20:34.012149418 +0000 UTC m=+8867.894775878" observedRunningTime="2025-10-05 09:20:34.677349261 +0000 UTC m=+8868.559975741" watchObservedRunningTime="2025-10-05 09:20:34.689189132 +0000 UTC m=+8868.571815592" Oct 05 09:20:36 crc kubenswrapper[4935]: I1005 09:20:36.684522 4935 generic.go:334] "Generic (PLEG): container finished" podID="43c59bed-43ce-4cef-9d35-45e7496c92f3" containerID="1460fd0eae980d2b38f57f12269f65be81faa7b71c9a90b92dd85028e4a00843" exitCode=0 Oct 05 09:20:36 crc kubenswrapper[4935]: I1005 09:20:36.684759 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-f2sdw" event={"ID":"43c59bed-43ce-4cef-9d35-45e7496c92f3","Type":"ContainerDied","Data":"1460fd0eae980d2b38f57f12269f65be81faa7b71c9a90b92dd85028e4a00843"} Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.167913 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.240049 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-inventory\") pod \"43c59bed-43ce-4cef-9d35-45e7496c92f3\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.240939 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-ssh-key\") pod \"43c59bed-43ce-4cef-9d35-45e7496c92f3\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.241070 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6qzq\" (UniqueName: \"kubernetes.io/projected/43c59bed-43ce-4cef-9d35-45e7496c92f3-kube-api-access-t6qzq\") pod \"43c59bed-43ce-4cef-9d35-45e7496c92f3\" (UID: \"43c59bed-43ce-4cef-9d35-45e7496c92f3\") " Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.245482 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c59bed-43ce-4cef-9d35-45e7496c92f3-kube-api-access-t6qzq" (OuterVolumeSpecName: "kube-api-access-t6qzq") pod "43c59bed-43ce-4cef-9d35-45e7496c92f3" (UID: "43c59bed-43ce-4cef-9d35-45e7496c92f3"). InnerVolumeSpecName "kube-api-access-t6qzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.272058 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43c59bed-43ce-4cef-9d35-45e7496c92f3" (UID: "43c59bed-43ce-4cef-9d35-45e7496c92f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.287903 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-inventory" (OuterVolumeSpecName: "inventory") pod "43c59bed-43ce-4cef-9d35-45e7496c92f3" (UID: "43c59bed-43ce-4cef-9d35-45e7496c92f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.344053 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.344416 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c59bed-43ce-4cef-9d35-45e7496c92f3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.344430 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6qzq\" (UniqueName: \"kubernetes.io/projected/43c59bed-43ce-4cef-9d35-45e7496c92f3-kube-api-access-t6qzq\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.707564 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-f2sdw" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.707613 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-f2sdw" event={"ID":"43c59bed-43ce-4cef-9d35-45e7496c92f3","Type":"ContainerDied","Data":"99071115e55e6ef7625bc71950dde5fb7cdef9b8036536a15f07c810221bace2"} Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.707674 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99071115e55e6ef7625bc71950dde5fb7cdef9b8036536a15f07c810221bace2" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.796161 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-dt6wh"] Oct 05 09:20:38 crc kubenswrapper[4935]: E1005 09:20:38.796753 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c59bed-43ce-4cef-9d35-45e7496c92f3" containerName="run-os-openstack-openstack-networker" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.796778 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c59bed-43ce-4cef-9d35-45e7496c92f3" containerName="run-os-openstack-openstack-networker" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.797175 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c59bed-43ce-4cef-9d35-45e7496c92f3" containerName="run-os-openstack-openstack-networker" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.798856 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-dt6wh"] Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.799088 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.801941 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.855069 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-inventory\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.855304 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2hnj\" (UniqueName: \"kubernetes.io/projected/8871439a-8466-4285-9872-747cfb0eb998-kube-api-access-j2hnj\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.856381 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-ssh-key\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.958579 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-inventory\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.958699 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2hnj\" (UniqueName: \"kubernetes.io/projected/8871439a-8466-4285-9872-747cfb0eb998-kube-api-access-j2hnj\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.958830 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-ssh-key\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.963710 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-inventory\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.964791 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-ssh-key\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:38 crc kubenswrapper[4935]: I1005 09:20:38.982541 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2hnj\" (UniqueName: \"kubernetes.io/projected/8871439a-8466-4285-9872-747cfb0eb998-kube-api-access-j2hnj\") pod \"reboot-os-openstack-openstack-networker-dt6wh\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:39 crc kubenswrapper[4935]: I1005 09:20:39.125546 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:39 crc kubenswrapper[4935]: I1005 09:20:39.722720 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-dt6wh"] Oct 05 09:20:40 crc kubenswrapper[4935]: I1005 09:20:40.726839 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" event={"ID":"8871439a-8466-4285-9872-747cfb0eb998","Type":"ContainerStarted","Data":"5ca3937642bd1efe50951e5d2d9e185bf0b6f78c917a361b6b144d0d2f5c8b23"} Oct 05 09:20:40 crc kubenswrapper[4935]: I1005 09:20:40.727170 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" event={"ID":"8871439a-8466-4285-9872-747cfb0eb998","Type":"ContainerStarted","Data":"f1c3ab1f17f8ccde911229a35a07dc08e35c62e184867ba2c4aa3d2a2d3585d8"} Oct 05 09:20:40 crc kubenswrapper[4935]: I1005 09:20:40.750925 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" podStartSLOduration=2.306817487 podStartE2EDuration="2.750909509s" podCreationTimestamp="2025-10-05 09:20:38 +0000 UTC" firstStartedPulling="2025-10-05 09:20:39.72976148 +0000 UTC m=+8873.612387940" lastFinishedPulling="2025-10-05 09:20:40.173853482 +0000 UTC m=+8874.056479962" observedRunningTime="2025-10-05 09:20:40.745409644 +0000 UTC m=+8874.628036104" watchObservedRunningTime="2025-10-05 09:20:40.750909509 +0000 UTC m=+8874.633535979" Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.289642 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.290530 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.290613 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.292079 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"989e4812ce9b8760f616df4a134bbde12de0cc616591fdb55b6764cc85626afe"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.292241 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://989e4812ce9b8760f616df4a134bbde12de0cc616591fdb55b6764cc85626afe" gracePeriod=600 Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.788069 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="989e4812ce9b8760f616df4a134bbde12de0cc616591fdb55b6764cc85626afe" exitCode=0 Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.789070 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"989e4812ce9b8760f616df4a134bbde12de0cc616591fdb55b6764cc85626afe"} Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.789113 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b"} Oct 05 09:20:44 crc kubenswrapper[4935]: I1005 09:20:44.789137 4935 scope.go:117] "RemoveContainer" containerID="3a9ea88c5a430676c326be4db52802875642f08d77432691a366a6be0eb3ee46" Oct 05 09:20:50 crc kubenswrapper[4935]: E1005 09:20:50.766252 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06081368_97c9_41f4_9fed_75c593994b11.slice/crio-conmon-f7ea98626907a5b0aacae9bfd4e810283589bda010a88ece9a970d14d9551604.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06081368_97c9_41f4_9fed_75c593994b11.slice/crio-f7ea98626907a5b0aacae9bfd4e810283589bda010a88ece9a970d14d9551604.scope\": RecentStats: unable to find data in memory cache]" Oct 05 09:20:50 crc kubenswrapper[4935]: I1005 09:20:50.862480 4935 generic.go:334] "Generic (PLEG): container finished" podID="06081368-97c9-41f4-9fed-75c593994b11" containerID="f7ea98626907a5b0aacae9bfd4e810283589bda010a88ece9a970d14d9551604" exitCode=0 Oct 05 09:20:50 crc kubenswrapper[4935]: I1005 09:20:50.862521 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-jn258" event={"ID":"06081368-97c9-41f4-9fed-75c593994b11","Type":"ContainerDied","Data":"f7ea98626907a5b0aacae9bfd4e810283589bda010a88ece9a970d14d9551604"} Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.478807 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.541176 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-1\") pod \"06081368-97c9-41f4-9fed-75c593994b11\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.541263 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ceph\") pod \"06081368-97c9-41f4-9fed-75c593994b11\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.541293 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-0\") pod \"06081368-97c9-41f4-9fed-75c593994b11\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.541346 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-networker\") pod \"06081368-97c9-41f4-9fed-75c593994b11\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.541476 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-cell1\") pod \"06081368-97c9-41f4-9fed-75c593994b11\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.541508 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnpzf\" (UniqueName: \"kubernetes.io/projected/06081368-97c9-41f4-9fed-75c593994b11-kube-api-access-nnpzf\") pod \"06081368-97c9-41f4-9fed-75c593994b11\" (UID: \"06081368-97c9-41f4-9fed-75c593994b11\") " Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.563165 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06081368-97c9-41f4-9fed-75c593994b11-kube-api-access-nnpzf" (OuterVolumeSpecName: "kube-api-access-nnpzf") pod "06081368-97c9-41f4-9fed-75c593994b11" (UID: "06081368-97c9-41f4-9fed-75c593994b11"). InnerVolumeSpecName "kube-api-access-nnpzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.563285 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ceph" (OuterVolumeSpecName: "ceph") pod "06081368-97c9-41f4-9fed-75c593994b11" (UID: "06081368-97c9-41f4-9fed-75c593994b11"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.569958 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "06081368-97c9-41f4-9fed-75c593994b11" (UID: "06081368-97c9-41f4-9fed-75c593994b11"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.570592 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-1" (OuterVolumeSpecName: "inventory-1") pod "06081368-97c9-41f4-9fed-75c593994b11" (UID: "06081368-97c9-41f4-9fed-75c593994b11"). InnerVolumeSpecName "inventory-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.572440 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "06081368-97c9-41f4-9fed-75c593994b11" (UID: "06081368-97c9-41f4-9fed-75c593994b11"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.580223 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "06081368-97c9-41f4-9fed-75c593994b11" (UID: "06081368-97c9-41f4-9fed-75c593994b11"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.646019 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.646059 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnpzf\" (UniqueName: \"kubernetes.io/projected/06081368-97c9-41f4-9fed-75c593994b11-kube-api-access-nnpzf\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.646073 4935 reconciler_common.go:293] "Volume detached for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.646086 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.646098 4935 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.646109 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06081368-97c9-41f4-9fed-75c593994b11-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.884043 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-jn258" event={"ID":"06081368-97c9-41f4-9fed-75c593994b11","Type":"ContainerDied","Data":"0b91eb2445ae6e448d0c6287973a81adab6cbc2f0dcd54b6f54dc79bc6e17164"} Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.884093 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b91eb2445ae6e448d0c6287973a81adab6cbc2f0dcd54b6f54dc79bc6e17164" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.884121 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-jn258" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.970826 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tkmbh"] Oct 05 09:20:52 crc kubenswrapper[4935]: E1005 09:20:52.971301 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06081368-97c9-41f4-9fed-75c593994b11" containerName="ssh-known-hosts-openstack" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.971326 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="06081368-97c9-41f4-9fed-75c593994b11" containerName="ssh-known-hosts-openstack" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.971596 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="06081368-97c9-41f4-9fed-75c593994b11" containerName="ssh-known-hosts-openstack" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.972536 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.978227 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.980088 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:20:52 crc kubenswrapper[4935]: I1005 09:20:52.983164 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tkmbh"] Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.053656 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drlbk\" (UniqueName: \"kubernetes.io/projected/0982ce01-4f77-4cf8-a024-bdb05e7016f7-kube-api-access-drlbk\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.054127 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-inventory\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.054211 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ssh-key\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.054237 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ceph\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.155372 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ceph\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.155738 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drlbk\" (UniqueName: \"kubernetes.io/projected/0982ce01-4f77-4cf8-a024-bdb05e7016f7-kube-api-access-drlbk\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.155954 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-inventory\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.156058 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ssh-key\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.161639 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-inventory\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.161799 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ssh-key\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.165367 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ceph\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.190922 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drlbk\" (UniqueName: \"kubernetes.io/projected/0982ce01-4f77-4cf8-a024-bdb05e7016f7-kube-api-access-drlbk\") pod \"run-os-openstack-openstack-cell1-tkmbh\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.301480 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:20:53 crc kubenswrapper[4935]: I1005 09:20:53.895399 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tkmbh"] Oct 05 09:20:54 crc kubenswrapper[4935]: W1005 09:20:54.571224 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0982ce01_4f77_4cf8_a024_bdb05e7016f7.slice/crio-e70b45c8b177acb9d135b2833bbdba20011bea2a4a290f5c0ec0daa6b14b860a WatchSource:0}: Error finding container e70b45c8b177acb9d135b2833bbdba20011bea2a4a290f5c0ec0daa6b14b860a: Status 404 returned error can't find the container with id e70b45c8b177acb9d135b2833bbdba20011bea2a4a290f5c0ec0daa6b14b860a Oct 05 09:20:54 crc kubenswrapper[4935]: I1005 09:20:54.917514 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" event={"ID":"0982ce01-4f77-4cf8-a024-bdb05e7016f7","Type":"ContainerStarted","Data":"e70b45c8b177acb9d135b2833bbdba20011bea2a4a290f5c0ec0daa6b14b860a"} Oct 05 09:20:55 crc kubenswrapper[4935]: I1005 09:20:55.928973 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" event={"ID":"0982ce01-4f77-4cf8-a024-bdb05e7016f7","Type":"ContainerStarted","Data":"ea295b1a12c5c2240fe28eb4f41d33e85b9cb0858ecde51f52bb0329d229d147"} Oct 05 09:20:55 crc kubenswrapper[4935]: I1005 09:20:55.952418 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" podStartSLOduration=3.5778426899999998 podStartE2EDuration="3.952397464s" podCreationTimestamp="2025-10-05 09:20:52 +0000 UTC" firstStartedPulling="2025-10-05 09:20:54.575852355 +0000 UTC m=+8888.458478845" lastFinishedPulling="2025-10-05 09:20:54.950407159 +0000 UTC m=+8888.833033619" observedRunningTime="2025-10-05 09:20:55.944120767 +0000 UTC m=+8889.826747227" watchObservedRunningTime="2025-10-05 09:20:55.952397464 +0000 UTC m=+8889.835023914" Oct 05 09:20:57 crc kubenswrapper[4935]: I1005 09:20:57.950750 4935 generic.go:334] "Generic (PLEG): container finished" podID="8871439a-8466-4285-9872-747cfb0eb998" containerID="5ca3937642bd1efe50951e5d2d9e185bf0b6f78c917a361b6b144d0d2f5c8b23" exitCode=0 Oct 05 09:20:57 crc kubenswrapper[4935]: I1005 09:20:57.951022 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" event={"ID":"8871439a-8466-4285-9872-747cfb0eb998","Type":"ContainerDied","Data":"5ca3937642bd1efe50951e5d2d9e185bf0b6f78c917a361b6b144d0d2f5c8b23"} Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.486199 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.595687 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-ssh-key\") pod \"8871439a-8466-4285-9872-747cfb0eb998\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.595739 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-inventory\") pod \"8871439a-8466-4285-9872-747cfb0eb998\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.595781 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2hnj\" (UniqueName: \"kubernetes.io/projected/8871439a-8466-4285-9872-747cfb0eb998-kube-api-access-j2hnj\") pod \"8871439a-8466-4285-9872-747cfb0eb998\" (UID: \"8871439a-8466-4285-9872-747cfb0eb998\") " Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.601277 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8871439a-8466-4285-9872-747cfb0eb998-kube-api-access-j2hnj" (OuterVolumeSpecName: "kube-api-access-j2hnj") pod "8871439a-8466-4285-9872-747cfb0eb998" (UID: "8871439a-8466-4285-9872-747cfb0eb998"). InnerVolumeSpecName "kube-api-access-j2hnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.624708 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-inventory" (OuterVolumeSpecName: "inventory") pod "8871439a-8466-4285-9872-747cfb0eb998" (UID: "8871439a-8466-4285-9872-747cfb0eb998"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.627853 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8871439a-8466-4285-9872-747cfb0eb998" (UID: "8871439a-8466-4285-9872-747cfb0eb998"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.698686 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.699015 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8871439a-8466-4285-9872-747cfb0eb998-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.699030 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2hnj\" (UniqueName: \"kubernetes.io/projected/8871439a-8466-4285-9872-747cfb0eb998-kube-api-access-j2hnj\") on node \"crc\" DevicePath \"\"" Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.973587 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" event={"ID":"8871439a-8466-4285-9872-747cfb0eb998","Type":"ContainerDied","Data":"f1c3ab1f17f8ccde911229a35a07dc08e35c62e184867ba2c4aa3d2a2d3585d8"} Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.973820 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1c3ab1f17f8ccde911229a35a07dc08e35c62e184867ba2c4aa3d2a2d3585d8" Oct 05 09:20:59 crc kubenswrapper[4935]: I1005 09:20:59.973814 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-dt6wh" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.054160 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-networker-5rb6l"] Oct 05 09:21:00 crc kubenswrapper[4935]: E1005 09:21:00.054703 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8871439a-8466-4285-9872-747cfb0eb998" containerName="reboot-os-openstack-openstack-networker" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.054729 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8871439a-8466-4285-9872-747cfb0eb998" containerName="reboot-os-openstack-openstack-networker" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.055022 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8871439a-8466-4285-9872-747cfb0eb998" containerName="reboot-os-openstack-openstack-networker" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.055959 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.060370 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.060413 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.081284 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-5rb6l"] Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.112210 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ssh-key\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.112254 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.112346 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.112429 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-inventory\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.112847 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kdbn\" (UniqueName: \"kubernetes.io/projected/6efd4306-da0b-4596-a2b8-20614f208aa7-kube-api-access-7kdbn\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.113246 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.216022 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kdbn\" (UniqueName: \"kubernetes.io/projected/6efd4306-da0b-4596-a2b8-20614f208aa7-kube-api-access-7kdbn\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.216297 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.216414 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ssh-key\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.216458 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.216553 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.216594 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-inventory\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.221204 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ssh-key\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.224416 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-inventory\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.225759 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.226265 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.227294 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.247306 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kdbn\" (UniqueName: \"kubernetes.io/projected/6efd4306-da0b-4596-a2b8-20614f208aa7-kube-api-access-7kdbn\") pod \"install-certs-openstack-openstack-networker-5rb6l\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.386498 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.840529 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-5rb6l"] Oct 05 09:21:00 crc kubenswrapper[4935]: W1005 09:21:00.843021 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6efd4306_da0b_4596_a2b8_20614f208aa7.slice/crio-5bfb3d2dbac6f120897e45eded9e66335e8de8b942c920decd128f471c94105a WatchSource:0}: Error finding container 5bfb3d2dbac6f120897e45eded9e66335e8de8b942c920decd128f471c94105a: Status 404 returned error can't find the container with id 5bfb3d2dbac6f120897e45eded9e66335e8de8b942c920decd128f471c94105a Oct 05 09:21:00 crc kubenswrapper[4935]: I1005 09:21:00.982101 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" event={"ID":"6efd4306-da0b-4596-a2b8-20614f208aa7","Type":"ContainerStarted","Data":"5bfb3d2dbac6f120897e45eded9e66335e8de8b942c920decd128f471c94105a"} Oct 05 09:21:01 crc kubenswrapper[4935]: I1005 09:21:01.994995 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" event={"ID":"6efd4306-da0b-4596-a2b8-20614f208aa7","Type":"ContainerStarted","Data":"74d3719b1d9d5c0a2770777032996b030f7beef143553bef3e503f75ebcb314a"} Oct 05 09:21:02 crc kubenswrapper[4935]: I1005 09:21:02.024993 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" podStartSLOduration=1.629490304 podStartE2EDuration="2.024970607s" podCreationTimestamp="2025-10-05 09:21:00 +0000 UTC" firstStartedPulling="2025-10-05 09:21:00.847017458 +0000 UTC m=+8894.729643918" lastFinishedPulling="2025-10-05 09:21:01.242497761 +0000 UTC m=+8895.125124221" observedRunningTime="2025-10-05 09:21:02.022233725 +0000 UTC m=+8895.904860335" watchObservedRunningTime="2025-10-05 09:21:02.024970607 +0000 UTC m=+8895.907597077" Oct 05 09:21:05 crc kubenswrapper[4935]: I1005 09:21:05.042578 4935 generic.go:334] "Generic (PLEG): container finished" podID="0982ce01-4f77-4cf8-a024-bdb05e7016f7" containerID="ea295b1a12c5c2240fe28eb4f41d33e85b9cb0858ecde51f52bb0329d229d147" exitCode=0 Oct 05 09:21:05 crc kubenswrapper[4935]: I1005 09:21:05.042657 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" event={"ID":"0982ce01-4f77-4cf8-a024-bdb05e7016f7","Type":"ContainerDied","Data":"ea295b1a12c5c2240fe28eb4f41d33e85b9cb0858ecde51f52bb0329d229d147"} Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.669123 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.775298 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drlbk\" (UniqueName: \"kubernetes.io/projected/0982ce01-4f77-4cf8-a024-bdb05e7016f7-kube-api-access-drlbk\") pod \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.775373 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ssh-key\") pod \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.775446 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-inventory\") pod \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.775480 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ceph\") pod \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\" (UID: \"0982ce01-4f77-4cf8-a024-bdb05e7016f7\") " Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.788583 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ceph" (OuterVolumeSpecName: "ceph") pod "0982ce01-4f77-4cf8-a024-bdb05e7016f7" (UID: "0982ce01-4f77-4cf8-a024-bdb05e7016f7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.796058 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0982ce01-4f77-4cf8-a024-bdb05e7016f7-kube-api-access-drlbk" (OuterVolumeSpecName: "kube-api-access-drlbk") pod "0982ce01-4f77-4cf8-a024-bdb05e7016f7" (UID: "0982ce01-4f77-4cf8-a024-bdb05e7016f7"). InnerVolumeSpecName "kube-api-access-drlbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.816073 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-inventory" (OuterVolumeSpecName: "inventory") pod "0982ce01-4f77-4cf8-a024-bdb05e7016f7" (UID: "0982ce01-4f77-4cf8-a024-bdb05e7016f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.819482 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0982ce01-4f77-4cf8-a024-bdb05e7016f7" (UID: "0982ce01-4f77-4cf8-a024-bdb05e7016f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.879138 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.879179 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.879194 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drlbk\" (UniqueName: \"kubernetes.io/projected/0982ce01-4f77-4cf8-a024-bdb05e7016f7-kube-api-access-drlbk\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:06 crc kubenswrapper[4935]: I1005 09:21:06.879207 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0982ce01-4f77-4cf8-a024-bdb05e7016f7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.067204 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" event={"ID":"0982ce01-4f77-4cf8-a024-bdb05e7016f7","Type":"ContainerDied","Data":"e70b45c8b177acb9d135b2833bbdba20011bea2a4a290f5c0ec0daa6b14b860a"} Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.067475 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e70b45c8b177acb9d135b2833bbdba20011bea2a4a290f5c0ec0daa6b14b860a" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.067397 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tkmbh" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.154444 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-dd458"] Oct 05 09:21:07 crc kubenswrapper[4935]: E1005 09:21:07.154918 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0982ce01-4f77-4cf8-a024-bdb05e7016f7" containerName="run-os-openstack-openstack-cell1" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.154937 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0982ce01-4f77-4cf8-a024-bdb05e7016f7" containerName="run-os-openstack-openstack-cell1" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.155168 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0982ce01-4f77-4cf8-a024-bdb05e7016f7" containerName="run-os-openstack-openstack-cell1" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.155926 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.158867 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.165277 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.170089 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-dd458"] Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.309771 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-inventory\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.310014 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xf7f\" (UniqueName: \"kubernetes.io/projected/db79fc8f-401b-426f-9f03-aada57d9d905-kube-api-access-8xf7f\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.310142 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ceph\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.310208 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.413093 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-inventory\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.413562 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xf7f\" (UniqueName: \"kubernetes.io/projected/db79fc8f-401b-426f-9f03-aada57d9d905-kube-api-access-8xf7f\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.414062 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ceph\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.414086 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.417550 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.417807 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-inventory\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.428390 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ceph\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.430088 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xf7f\" (UniqueName: \"kubernetes.io/projected/db79fc8f-401b-426f-9f03-aada57d9d905-kube-api-access-8xf7f\") pod \"reboot-os-openstack-openstack-cell1-dd458\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:07 crc kubenswrapper[4935]: I1005 09:21:07.520495 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:08 crc kubenswrapper[4935]: W1005 09:21:08.100176 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb79fc8f_401b_426f_9f03_aada57d9d905.slice/crio-6cbe8cfd25eafa61274a2367846c930f0c20c4091a34f0a8c31c1624dd1a4c02 WatchSource:0}: Error finding container 6cbe8cfd25eafa61274a2367846c930f0c20c4091a34f0a8c31c1624dd1a4c02: Status 404 returned error can't find the container with id 6cbe8cfd25eafa61274a2367846c930f0c20c4091a34f0a8c31c1624dd1a4c02 Oct 05 09:21:08 crc kubenswrapper[4935]: I1005 09:21:08.100321 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-dd458"] Oct 05 09:21:09 crc kubenswrapper[4935]: I1005 09:21:09.088054 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" event={"ID":"db79fc8f-401b-426f-9f03-aada57d9d905","Type":"ContainerStarted","Data":"cb873d493b131b10628fd875548ab6781df4fc091fde299febd964e2d28aeb35"} Oct 05 09:21:09 crc kubenswrapper[4935]: I1005 09:21:09.088682 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" event={"ID":"db79fc8f-401b-426f-9f03-aada57d9d905","Type":"ContainerStarted","Data":"6cbe8cfd25eafa61274a2367846c930f0c20c4091a34f0a8c31c1624dd1a4c02"} Oct 05 09:21:09 crc kubenswrapper[4935]: I1005 09:21:09.124341 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" podStartSLOduration=1.7081247290000001 podStartE2EDuration="2.124312577s" podCreationTimestamp="2025-10-05 09:21:07 +0000 UTC" firstStartedPulling="2025-10-05 09:21:08.104637977 +0000 UTC m=+8901.987264437" lastFinishedPulling="2025-10-05 09:21:08.520825825 +0000 UTC m=+8902.403452285" observedRunningTime="2025-10-05 09:21:09.11338786 +0000 UTC m=+8902.996014370" watchObservedRunningTime="2025-10-05 09:21:09.124312577 +0000 UTC m=+8903.006939067" Oct 05 09:21:13 crc kubenswrapper[4935]: I1005 09:21:13.132456 4935 generic.go:334] "Generic (PLEG): container finished" podID="6efd4306-da0b-4596-a2b8-20614f208aa7" containerID="74d3719b1d9d5c0a2770777032996b030f7beef143553bef3e503f75ebcb314a" exitCode=0 Oct 05 09:21:13 crc kubenswrapper[4935]: I1005 09:21:13.132551 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" event={"ID":"6efd4306-da0b-4596-a2b8-20614f208aa7","Type":"ContainerDied","Data":"74d3719b1d9d5c0a2770777032996b030f7beef143553bef3e503f75ebcb314a"} Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.637834 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.790113 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-neutron-metadata-combined-ca-bundle\") pod \"6efd4306-da0b-4596-a2b8-20614f208aa7\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.791156 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ovn-combined-ca-bundle\") pod \"6efd4306-da0b-4596-a2b8-20614f208aa7\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.791188 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ssh-key\") pod \"6efd4306-da0b-4596-a2b8-20614f208aa7\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.791217 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-bootstrap-combined-ca-bundle\") pod \"6efd4306-da0b-4596-a2b8-20614f208aa7\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.791237 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kdbn\" (UniqueName: \"kubernetes.io/projected/6efd4306-da0b-4596-a2b8-20614f208aa7-kube-api-access-7kdbn\") pod \"6efd4306-da0b-4596-a2b8-20614f208aa7\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.791286 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-inventory\") pod \"6efd4306-da0b-4596-a2b8-20614f208aa7\" (UID: \"6efd4306-da0b-4596-a2b8-20614f208aa7\") " Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.801690 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6efd4306-da0b-4596-a2b8-20614f208aa7" (UID: "6efd4306-da0b-4596-a2b8-20614f208aa7"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.801812 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6efd4306-da0b-4596-a2b8-20614f208aa7" (UID: "6efd4306-da0b-4596-a2b8-20614f208aa7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.802902 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6efd4306-da0b-4596-a2b8-20614f208aa7-kube-api-access-7kdbn" (OuterVolumeSpecName: "kube-api-access-7kdbn") pod "6efd4306-da0b-4596-a2b8-20614f208aa7" (UID: "6efd4306-da0b-4596-a2b8-20614f208aa7"). InnerVolumeSpecName "kube-api-access-7kdbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.804381 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6efd4306-da0b-4596-a2b8-20614f208aa7" (UID: "6efd4306-da0b-4596-a2b8-20614f208aa7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.835629 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6efd4306-da0b-4596-a2b8-20614f208aa7" (UID: "6efd4306-da0b-4596-a2b8-20614f208aa7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.840268 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-inventory" (OuterVolumeSpecName: "inventory") pod "6efd4306-da0b-4596-a2b8-20614f208aa7" (UID: "6efd4306-da0b-4596-a2b8-20614f208aa7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.894444 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.894480 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.894493 4935 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.894505 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kdbn\" (UniqueName: \"kubernetes.io/projected/6efd4306-da0b-4596-a2b8-20614f208aa7-kube-api-access-7kdbn\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.894519 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:14 crc kubenswrapper[4935]: I1005 09:21:14.894530 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efd4306-da0b-4596-a2b8-20614f208aa7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.153290 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" event={"ID":"6efd4306-da0b-4596-a2b8-20614f208aa7","Type":"ContainerDied","Data":"5bfb3d2dbac6f120897e45eded9e66335e8de8b942c920decd128f471c94105a"} Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.153560 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bfb3d2dbac6f120897e45eded9e66335e8de8b942c920decd128f471c94105a" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.153352 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-5rb6l" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.226734 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-networker-zgwf7"] Oct 05 09:21:15 crc kubenswrapper[4935]: E1005 09:21:15.227170 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6efd4306-da0b-4596-a2b8-20614f208aa7" containerName="install-certs-openstack-openstack-networker" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.227187 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6efd4306-da0b-4596-a2b8-20614f208aa7" containerName="install-certs-openstack-openstack-networker" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.227357 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6efd4306-da0b-4596-a2b8-20614f208aa7" containerName="install-certs-openstack-openstack-networker" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.228068 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.235659 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.235905 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.236424 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.256590 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-zgwf7"] Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.301346 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z4zz\" (UniqueName: \"kubernetes.io/projected/7b5b8e58-6662-4dac-8700-f324b232653b-kube-api-access-2z4zz\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.301420 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ssh-key\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.301486 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.301550 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-inventory\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.301593 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7b5b8e58-6662-4dac-8700-f324b232653b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.403395 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ssh-key\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.403525 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.403635 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-inventory\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.403684 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7b5b8e58-6662-4dac-8700-f324b232653b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.403729 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z4zz\" (UniqueName: \"kubernetes.io/projected/7b5b8e58-6662-4dac-8700-f324b232653b-kube-api-access-2z4zz\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.404861 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7b5b8e58-6662-4dac-8700-f324b232653b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.407539 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ssh-key\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.407880 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-inventory\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.410388 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.419803 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z4zz\" (UniqueName: \"kubernetes.io/projected/7b5b8e58-6662-4dac-8700-f324b232653b-kube-api-access-2z4zz\") pod \"ovn-openstack-openstack-networker-zgwf7\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:15 crc kubenswrapper[4935]: I1005 09:21:15.550225 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:21:16 crc kubenswrapper[4935]: I1005 09:21:16.099335 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-zgwf7"] Oct 05 09:21:16 crc kubenswrapper[4935]: W1005 09:21:16.370951 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b5b8e58_6662_4dac_8700_f324b232653b.slice/crio-eed14309720934501f514df0e81f23e2efd10b357e361675a37dbb3cc7ed3420 WatchSource:0}: Error finding container eed14309720934501f514df0e81f23e2efd10b357e361675a37dbb3cc7ed3420: Status 404 returned error can't find the container with id eed14309720934501f514df0e81f23e2efd10b357e361675a37dbb3cc7ed3420 Oct 05 09:21:17 crc kubenswrapper[4935]: I1005 09:21:17.175013 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-zgwf7" event={"ID":"7b5b8e58-6662-4dac-8700-f324b232653b","Type":"ContainerStarted","Data":"0196ac346180e43dcc2a063ca1d829ec45adc7c5f171f61173b8e04d1e1d31a2"} Oct 05 09:21:17 crc kubenswrapper[4935]: I1005 09:21:17.175439 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-zgwf7" event={"ID":"7b5b8e58-6662-4dac-8700-f324b232653b","Type":"ContainerStarted","Data":"eed14309720934501f514df0e81f23e2efd10b357e361675a37dbb3cc7ed3420"} Oct 05 09:21:17 crc kubenswrapper[4935]: I1005 09:21:17.198921 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-networker-zgwf7" podStartSLOduration=1.77663193 podStartE2EDuration="2.198886538s" podCreationTimestamp="2025-10-05 09:21:15 +0000 UTC" firstStartedPulling="2025-10-05 09:21:16.373955376 +0000 UTC m=+8910.256581836" lastFinishedPulling="2025-10-05 09:21:16.796209984 +0000 UTC m=+8910.678836444" observedRunningTime="2025-10-05 09:21:17.187467388 +0000 UTC m=+8911.070093848" watchObservedRunningTime="2025-10-05 09:21:17.198886538 +0000 UTC m=+8911.081512998" Oct 05 09:21:24 crc kubenswrapper[4935]: I1005 09:21:24.258157 4935 generic.go:334] "Generic (PLEG): container finished" podID="db79fc8f-401b-426f-9f03-aada57d9d905" containerID="cb873d493b131b10628fd875548ab6781df4fc091fde299febd964e2d28aeb35" exitCode=0 Oct 05 09:21:24 crc kubenswrapper[4935]: I1005 09:21:24.258237 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" event={"ID":"db79fc8f-401b-426f-9f03-aada57d9d905","Type":"ContainerDied","Data":"cb873d493b131b10628fd875548ab6781df4fc091fde299febd964e2d28aeb35"} Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.675718 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.847589 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ssh-key\") pod \"db79fc8f-401b-426f-9f03-aada57d9d905\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.848525 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xf7f\" (UniqueName: \"kubernetes.io/projected/db79fc8f-401b-426f-9f03-aada57d9d905-kube-api-access-8xf7f\") pod \"db79fc8f-401b-426f-9f03-aada57d9d905\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.848590 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ceph\") pod \"db79fc8f-401b-426f-9f03-aada57d9d905\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.848615 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-inventory\") pod \"db79fc8f-401b-426f-9f03-aada57d9d905\" (UID: \"db79fc8f-401b-426f-9f03-aada57d9d905\") " Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.853554 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db79fc8f-401b-426f-9f03-aada57d9d905-kube-api-access-8xf7f" (OuterVolumeSpecName: "kube-api-access-8xf7f") pod "db79fc8f-401b-426f-9f03-aada57d9d905" (UID: "db79fc8f-401b-426f-9f03-aada57d9d905"). InnerVolumeSpecName "kube-api-access-8xf7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.853869 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ceph" (OuterVolumeSpecName: "ceph") pod "db79fc8f-401b-426f-9f03-aada57d9d905" (UID: "db79fc8f-401b-426f-9f03-aada57d9d905"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.879499 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-inventory" (OuterVolumeSpecName: "inventory") pod "db79fc8f-401b-426f-9f03-aada57d9d905" (UID: "db79fc8f-401b-426f-9f03-aada57d9d905"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.882489 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db79fc8f-401b-426f-9f03-aada57d9d905" (UID: "db79fc8f-401b-426f-9f03-aada57d9d905"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.950873 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.950952 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.950975 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db79fc8f-401b-426f-9f03-aada57d9d905-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:25 crc kubenswrapper[4935]: I1005 09:21:25.950996 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xf7f\" (UniqueName: \"kubernetes.io/projected/db79fc8f-401b-426f-9f03-aada57d9d905-kube-api-access-8xf7f\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.282159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" event={"ID":"db79fc8f-401b-426f-9f03-aada57d9d905","Type":"ContainerDied","Data":"6cbe8cfd25eafa61274a2367846c930f0c20c4091a34f0a8c31c1624dd1a4c02"} Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.282492 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cbe8cfd25eafa61274a2367846c930f0c20c4091a34f0a8c31c1624dd1a4c02" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.282213 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-dd458" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.351191 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vpc8x"] Oct 05 09:21:26 crc kubenswrapper[4935]: E1005 09:21:26.351579 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db79fc8f-401b-426f-9f03-aada57d9d905" containerName="reboot-os-openstack-openstack-cell1" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.351597 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="db79fc8f-401b-426f-9f03-aada57d9d905" containerName="reboot-os-openstack-openstack-cell1" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.351802 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="db79fc8f-401b-426f-9f03-aada57d9d905" containerName="reboot-os-openstack-openstack-cell1" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.352484 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.354493 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.354863 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.369760 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vpc8x"] Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.460433 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.460692 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.460836 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.460904 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ceph\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.460931 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.460969 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.460998 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-inventory\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.461079 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94jm5\" (UniqueName: \"kubernetes.io/projected/f9f3edd1-359c-4a45-bcdb-026066428d7f-kube-api-access-94jm5\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.461128 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.461145 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.461197 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.461228 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563249 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563310 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563362 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563397 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ceph\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563419 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563457 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563478 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-inventory\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563651 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94jm5\" (UniqueName: \"kubernetes.io/projected/f9f3edd1-359c-4a45-bcdb-026066428d7f-kube-api-access-94jm5\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563685 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563703 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563733 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.563766 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.567826 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ceph\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.568481 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-inventory\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.568651 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.569077 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.570830 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.570858 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.571648 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.572627 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.573466 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.573687 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.582565 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.583094 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94jm5\" (UniqueName: \"kubernetes.io/projected/f9f3edd1-359c-4a45-bcdb-026066428d7f-kube-api-access-94jm5\") pod \"install-certs-openstack-openstack-cell1-vpc8x\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:26 crc kubenswrapper[4935]: I1005 09:21:26.669238 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:27 crc kubenswrapper[4935]: I1005 09:21:27.988405 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vpc8x"] Oct 05 09:21:28 crc kubenswrapper[4935]: I1005 09:21:28.302482 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" event={"ID":"f9f3edd1-359c-4a45-bcdb-026066428d7f","Type":"ContainerStarted","Data":"dcde27f07de09d5b995b2690be5f9814afdd407aa9fe2d27e8358949a337f78b"} Oct 05 09:21:29 crc kubenswrapper[4935]: I1005 09:21:29.318531 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" event={"ID":"f9f3edd1-359c-4a45-bcdb-026066428d7f","Type":"ContainerStarted","Data":"3dd44265b39c4e6072a52f356dff74905fd5649f44b1567165940b333dfd3081"} Oct 05 09:21:29 crc kubenswrapper[4935]: I1005 09:21:29.338628 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" podStartSLOduration=2.915552853 podStartE2EDuration="3.338611223s" podCreationTimestamp="2025-10-05 09:21:26 +0000 UTC" firstStartedPulling="2025-10-05 09:21:28.001477859 +0000 UTC m=+8921.884104319" lastFinishedPulling="2025-10-05 09:21:28.424536229 +0000 UTC m=+8922.307162689" observedRunningTime="2025-10-05 09:21:29.335667176 +0000 UTC m=+8923.218293656" watchObservedRunningTime="2025-10-05 09:21:29.338611223 +0000 UTC m=+8923.221237683" Oct 05 09:21:50 crc kubenswrapper[4935]: I1005 09:21:50.535910 4935 generic.go:334] "Generic (PLEG): container finished" podID="f9f3edd1-359c-4a45-bcdb-026066428d7f" containerID="3dd44265b39c4e6072a52f356dff74905fd5649f44b1567165940b333dfd3081" exitCode=0 Oct 05 09:21:50 crc kubenswrapper[4935]: I1005 09:21:50.536017 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" event={"ID":"f9f3edd1-359c-4a45-bcdb-026066428d7f","Type":"ContainerDied","Data":"3dd44265b39c4e6072a52f356dff74905fd5649f44b1567165940b333dfd3081"} Oct 05 09:21:51 crc kubenswrapper[4935]: I1005 09:21:51.998784 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.125921 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-libvirt-combined-ca-bundle\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126308 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-telemetry-combined-ca-bundle\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126428 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-inventory\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126464 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94jm5\" (UniqueName: \"kubernetes.io/projected/f9f3edd1-359c-4a45-bcdb-026066428d7f-kube-api-access-94jm5\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126499 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-nova-combined-ca-bundle\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126559 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ssh-key\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126588 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-bootstrap-combined-ca-bundle\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126625 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-dhcp-combined-ca-bundle\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126650 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-sriov-combined-ca-bundle\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126698 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ceph\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126726 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-metadata-combined-ca-bundle\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.126752 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ovn-combined-ca-bundle\") pod \"f9f3edd1-359c-4a45-bcdb-026066428d7f\" (UID: \"f9f3edd1-359c-4a45-bcdb-026066428d7f\") " Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.133723 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.135763 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9f3edd1-359c-4a45-bcdb-026066428d7f-kube-api-access-94jm5" (OuterVolumeSpecName: "kube-api-access-94jm5") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "kube-api-access-94jm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.136267 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.138196 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ceph" (OuterVolumeSpecName: "ceph") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.138238 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.140152 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.140751 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.143110 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.144049 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.164396 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.167998 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.173860 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-inventory" (OuterVolumeSpecName: "inventory") pod "f9f3edd1-359c-4a45-bcdb-026066428d7f" (UID: "f9f3edd1-359c-4a45-bcdb-026066428d7f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234166 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234208 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234223 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234238 4935 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234250 4935 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234260 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234273 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94jm5\" (UniqueName: \"kubernetes.io/projected/f9f3edd1-359c-4a45-bcdb-026066428d7f-kube-api-access-94jm5\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234284 4935 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234294 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234305 4935 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234316 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.234327 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3edd1-359c-4a45-bcdb-026066428d7f-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.558165 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" event={"ID":"f9f3edd1-359c-4a45-bcdb-026066428d7f","Type":"ContainerDied","Data":"dcde27f07de09d5b995b2690be5f9814afdd407aa9fe2d27e8358949a337f78b"} Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.558217 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcde27f07de09d5b995b2690be5f9814afdd407aa9fe2d27e8358949a337f78b" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.558220 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vpc8x" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.654309 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-gt9nr"] Oct 05 09:21:52 crc kubenswrapper[4935]: E1005 09:21:52.654861 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f3edd1-359c-4a45-bcdb-026066428d7f" containerName="install-certs-openstack-openstack-cell1" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.654887 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f3edd1-359c-4a45-bcdb-026066428d7f" containerName="install-certs-openstack-openstack-cell1" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.655259 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f3edd1-359c-4a45-bcdb-026066428d7f" containerName="install-certs-openstack-openstack-cell1" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.656023 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.658224 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.658415 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.683110 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-gt9nr"] Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.844321 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ceph\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.844439 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.844531 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knwnm\" (UniqueName: \"kubernetes.io/projected/ecda3b70-058d-490d-a537-7511163dad56-kube-api-access-knwnm\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.844555 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-inventory\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.946969 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knwnm\" (UniqueName: \"kubernetes.io/projected/ecda3b70-058d-490d-a537-7511163dad56-kube-api-access-knwnm\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.947052 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-inventory\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.947209 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ceph\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.947370 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.951036 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.951195 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ceph\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.966625 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-inventory\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.969920 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knwnm\" (UniqueName: \"kubernetes.io/projected/ecda3b70-058d-490d-a537-7511163dad56-kube-api-access-knwnm\") pod \"ceph-client-openstack-openstack-cell1-gt9nr\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:52 crc kubenswrapper[4935]: I1005 09:21:52.981083 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:21:53 crc kubenswrapper[4935]: I1005 09:21:53.589905 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-gt9nr"] Oct 05 09:21:54 crc kubenswrapper[4935]: I1005 09:21:54.579847 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" event={"ID":"ecda3b70-058d-490d-a537-7511163dad56","Type":"ContainerStarted","Data":"782f312009d8424d6709a602a8ae23db1bded5cef151981c0c8d7356a7e11f00"} Oct 05 09:21:54 crc kubenswrapper[4935]: I1005 09:21:54.580351 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" event={"ID":"ecda3b70-058d-490d-a537-7511163dad56","Type":"ContainerStarted","Data":"f315a1f37b4658eeff2f902cbc220981edbdc9f2a0ec8597c74caf3803b18b2e"} Oct 05 09:21:54 crc kubenswrapper[4935]: I1005 09:21:54.605479 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" podStartSLOduration=1.965788259 podStartE2EDuration="2.605456552s" podCreationTimestamp="2025-10-05 09:21:52 +0000 UTC" firstStartedPulling="2025-10-05 09:21:53.591667827 +0000 UTC m=+8947.474294287" lastFinishedPulling="2025-10-05 09:21:54.23133612 +0000 UTC m=+8948.113962580" observedRunningTime="2025-10-05 09:21:54.595343136 +0000 UTC m=+8948.477969616" watchObservedRunningTime="2025-10-05 09:21:54.605456552 +0000 UTC m=+8948.488083012" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.713194 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-twxcz"] Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.715565 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.731229 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twxcz"] Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.808437 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b52pk\" (UniqueName: \"kubernetes.io/projected/a33183af-95fe-474c-ae5d-d288ab19c967-kube-api-access-b52pk\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.808517 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-catalog-content\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.808918 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-utilities\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.911464 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b52pk\" (UniqueName: \"kubernetes.io/projected/a33183af-95fe-474c-ae5d-d288ab19c967-kube-api-access-b52pk\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.911548 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-catalog-content\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.911635 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-utilities\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.912572 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-utilities\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.912599 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-catalog-content\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:55 crc kubenswrapper[4935]: I1005 09:21:55.932782 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b52pk\" (UniqueName: \"kubernetes.io/projected/a33183af-95fe-474c-ae5d-d288ab19c967-kube-api-access-b52pk\") pod \"community-operators-twxcz\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:56 crc kubenswrapper[4935]: I1005 09:21:56.075620 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:21:56 crc kubenswrapper[4935]: I1005 09:21:56.653125 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twxcz"] Oct 05 09:21:56 crc kubenswrapper[4935]: W1005 09:21:56.663782 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda33183af_95fe_474c_ae5d_d288ab19c967.slice/crio-6568aa07719027300c96660ec1142f945c49928bc44f0bfe3daf23c5ee00d613 WatchSource:0}: Error finding container 6568aa07719027300c96660ec1142f945c49928bc44f0bfe3daf23c5ee00d613: Status 404 returned error can't find the container with id 6568aa07719027300c96660ec1142f945c49928bc44f0bfe3daf23c5ee00d613 Oct 05 09:21:57 crc kubenswrapper[4935]: I1005 09:21:57.621004 4935 generic.go:334] "Generic (PLEG): container finished" podID="a33183af-95fe-474c-ae5d-d288ab19c967" containerID="7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb" exitCode=0 Oct 05 09:21:57 crc kubenswrapper[4935]: I1005 09:21:57.621087 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twxcz" event={"ID":"a33183af-95fe-474c-ae5d-d288ab19c967","Type":"ContainerDied","Data":"7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb"} Oct 05 09:21:57 crc kubenswrapper[4935]: I1005 09:21:57.621520 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twxcz" event={"ID":"a33183af-95fe-474c-ae5d-d288ab19c967","Type":"ContainerStarted","Data":"6568aa07719027300c96660ec1142f945c49928bc44f0bfe3daf23c5ee00d613"} Oct 05 09:21:58 crc kubenswrapper[4935]: I1005 09:21:58.637119 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twxcz" event={"ID":"a33183af-95fe-474c-ae5d-d288ab19c967","Type":"ContainerStarted","Data":"01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50"} Oct 05 09:22:00 crc kubenswrapper[4935]: I1005 09:22:00.663156 4935 generic.go:334] "Generic (PLEG): container finished" podID="a33183af-95fe-474c-ae5d-d288ab19c967" containerID="01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50" exitCode=0 Oct 05 09:22:00 crc kubenswrapper[4935]: I1005 09:22:00.663272 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twxcz" event={"ID":"a33183af-95fe-474c-ae5d-d288ab19c967","Type":"ContainerDied","Data":"01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50"} Oct 05 09:22:01 crc kubenswrapper[4935]: I1005 09:22:01.675057 4935 generic.go:334] "Generic (PLEG): container finished" podID="ecda3b70-058d-490d-a537-7511163dad56" containerID="782f312009d8424d6709a602a8ae23db1bded5cef151981c0c8d7356a7e11f00" exitCode=0 Oct 05 09:22:01 crc kubenswrapper[4935]: I1005 09:22:01.675150 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" event={"ID":"ecda3b70-058d-490d-a537-7511163dad56","Type":"ContainerDied","Data":"782f312009d8424d6709a602a8ae23db1bded5cef151981c0c8d7356a7e11f00"} Oct 05 09:22:01 crc kubenswrapper[4935]: I1005 09:22:01.682294 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twxcz" event={"ID":"a33183af-95fe-474c-ae5d-d288ab19c967","Type":"ContainerStarted","Data":"3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d"} Oct 05 09:22:01 crc kubenswrapper[4935]: I1005 09:22:01.725576 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-twxcz" podStartSLOduration=3.208113448 podStartE2EDuration="6.725553106s" podCreationTimestamp="2025-10-05 09:21:55 +0000 UTC" firstStartedPulling="2025-10-05 09:21:57.624589282 +0000 UTC m=+8951.507215752" lastFinishedPulling="2025-10-05 09:22:01.14202894 +0000 UTC m=+8955.024655410" observedRunningTime="2025-10-05 09:22:01.711660291 +0000 UTC m=+8955.594286751" watchObservedRunningTime="2025-10-05 09:22:01.725553106 +0000 UTC m=+8955.608179566" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.141357 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.187289 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knwnm\" (UniqueName: \"kubernetes.io/projected/ecda3b70-058d-490d-a537-7511163dad56-kube-api-access-knwnm\") pod \"ecda3b70-058d-490d-a537-7511163dad56\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.187331 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-inventory\") pod \"ecda3b70-058d-490d-a537-7511163dad56\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.187390 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ceph\") pod \"ecda3b70-058d-490d-a537-7511163dad56\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.187446 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ssh-key\") pod \"ecda3b70-058d-490d-a537-7511163dad56\" (UID: \"ecda3b70-058d-490d-a537-7511163dad56\") " Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.196144 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecda3b70-058d-490d-a537-7511163dad56-kube-api-access-knwnm" (OuterVolumeSpecName: "kube-api-access-knwnm") pod "ecda3b70-058d-490d-a537-7511163dad56" (UID: "ecda3b70-058d-490d-a537-7511163dad56"). InnerVolumeSpecName "kube-api-access-knwnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.217058 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ceph" (OuterVolumeSpecName: "ceph") pod "ecda3b70-058d-490d-a537-7511163dad56" (UID: "ecda3b70-058d-490d-a537-7511163dad56"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.221156 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-inventory" (OuterVolumeSpecName: "inventory") pod "ecda3b70-058d-490d-a537-7511163dad56" (UID: "ecda3b70-058d-490d-a537-7511163dad56"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.221828 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ecda3b70-058d-490d-a537-7511163dad56" (UID: "ecda3b70-058d-490d-a537-7511163dad56"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.289317 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knwnm\" (UniqueName: \"kubernetes.io/projected/ecda3b70-058d-490d-a537-7511163dad56-kube-api-access-knwnm\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.289352 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.289363 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.289373 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecda3b70-058d-490d-a537-7511163dad56-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.728512 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" event={"ID":"ecda3b70-058d-490d-a537-7511163dad56","Type":"ContainerDied","Data":"f315a1f37b4658eeff2f902cbc220981edbdc9f2a0ec8597c74caf3803b18b2e"} Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.728562 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f315a1f37b4658eeff2f902cbc220981edbdc9f2a0ec8597c74caf3803b18b2e" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.728587 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-gt9nr" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.801342 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-lfkb6"] Oct 05 09:22:03 crc kubenswrapper[4935]: E1005 09:22:03.802074 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecda3b70-058d-490d-a537-7511163dad56" containerName="ceph-client-openstack-openstack-cell1" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.802096 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecda3b70-058d-490d-a537-7511163dad56" containerName="ceph-client-openstack-openstack-cell1" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.802390 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecda3b70-058d-490d-a537-7511163dad56" containerName="ceph-client-openstack-openstack-cell1" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.803278 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.805367 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.805532 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.816190 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-lfkb6"] Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.900590 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.900639 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ssh-key\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.900667 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ceph\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.900805 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.900970 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnttb\" (UniqueName: \"kubernetes.io/projected/48e06b46-b928-49eb-a8b8-5b4b5d24925b-kube-api-access-jnttb\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:03 crc kubenswrapper[4935]: I1005 09:22:03.901102 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-inventory\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.002847 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.002929 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ssh-key\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.002973 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ceph\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.003019 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.003052 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnttb\" (UniqueName: \"kubernetes.io/projected/48e06b46-b928-49eb-a8b8-5b4b5d24925b-kube-api-access-jnttb\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.003131 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-inventory\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.003808 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.006700 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-inventory\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.006822 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ssh-key\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.007293 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ceph\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.011705 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.024910 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnttb\" (UniqueName: \"kubernetes.io/projected/48e06b46-b928-49eb-a8b8-5b4b5d24925b-kube-api-access-jnttb\") pod \"ovn-openstack-openstack-cell1-lfkb6\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.124605 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:22:04 crc kubenswrapper[4935]: W1005 09:22:04.745307 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48e06b46_b928_49eb_a8b8_5b4b5d24925b.slice/crio-6df551988ebc692e574aa0a89867e619148598e06d591451a1a76ff9b8b2db66 WatchSource:0}: Error finding container 6df551988ebc692e574aa0a89867e619148598e06d591451a1a76ff9b8b2db66: Status 404 returned error can't find the container with id 6df551988ebc692e574aa0a89867e619148598e06d591451a1a76ff9b8b2db66 Oct 05 09:22:04 crc kubenswrapper[4935]: I1005 09:22:04.752472 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-lfkb6"] Oct 05 09:22:05 crc kubenswrapper[4935]: I1005 09:22:05.752133 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" event={"ID":"48e06b46-b928-49eb-a8b8-5b4b5d24925b","Type":"ContainerStarted","Data":"e42a079019f667fadb51cf7abc27500a686649c32b6492553826a2c09c16e295"} Oct 05 09:22:05 crc kubenswrapper[4935]: I1005 09:22:05.752386 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" event={"ID":"48e06b46-b928-49eb-a8b8-5b4b5d24925b","Type":"ContainerStarted","Data":"6df551988ebc692e574aa0a89867e619148598e06d591451a1a76ff9b8b2db66"} Oct 05 09:22:05 crc kubenswrapper[4935]: I1005 09:22:05.777878 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" podStartSLOduration=2.343972007 podStartE2EDuration="2.777855961s" podCreationTimestamp="2025-10-05 09:22:03 +0000 UTC" firstStartedPulling="2025-10-05 09:22:04.747949853 +0000 UTC m=+8958.630576313" lastFinishedPulling="2025-10-05 09:22:05.181833807 +0000 UTC m=+8959.064460267" observedRunningTime="2025-10-05 09:22:05.770455647 +0000 UTC m=+8959.653082127" watchObservedRunningTime="2025-10-05 09:22:05.777855961 +0000 UTC m=+8959.660482431" Oct 05 09:22:06 crc kubenswrapper[4935]: I1005 09:22:06.076436 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:22:06 crc kubenswrapper[4935]: I1005 09:22:06.076493 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:22:06 crc kubenswrapper[4935]: I1005 09:22:06.730816 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:22:06 crc kubenswrapper[4935]: I1005 09:22:06.825934 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:22:06 crc kubenswrapper[4935]: I1005 09:22:06.967344 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twxcz"] Oct 05 09:22:08 crc kubenswrapper[4935]: I1005 09:22:08.821680 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-twxcz" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" containerName="registry-server" containerID="cri-o://3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d" gracePeriod=2 Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.333033 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.523428 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-utilities\") pod \"a33183af-95fe-474c-ae5d-d288ab19c967\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.523519 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-catalog-content\") pod \"a33183af-95fe-474c-ae5d-d288ab19c967\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.523660 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b52pk\" (UniqueName: \"kubernetes.io/projected/a33183af-95fe-474c-ae5d-d288ab19c967-kube-api-access-b52pk\") pod \"a33183af-95fe-474c-ae5d-d288ab19c967\" (UID: \"a33183af-95fe-474c-ae5d-d288ab19c967\") " Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.524431 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-utilities" (OuterVolumeSpecName: "utilities") pod "a33183af-95fe-474c-ae5d-d288ab19c967" (UID: "a33183af-95fe-474c-ae5d-d288ab19c967"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.531024 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a33183af-95fe-474c-ae5d-d288ab19c967-kube-api-access-b52pk" (OuterVolumeSpecName: "kube-api-access-b52pk") pod "a33183af-95fe-474c-ae5d-d288ab19c967" (UID: "a33183af-95fe-474c-ae5d-d288ab19c967"). InnerVolumeSpecName "kube-api-access-b52pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.581494 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a33183af-95fe-474c-ae5d-d288ab19c967" (UID: "a33183af-95fe-474c-ae5d-d288ab19c967"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.626484 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.626529 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33183af-95fe-474c-ae5d-d288ab19c967-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.626546 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b52pk\" (UniqueName: \"kubernetes.io/projected/a33183af-95fe-474c-ae5d-d288ab19c967-kube-api-access-b52pk\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.844272 4935 generic.go:334] "Generic (PLEG): container finished" podID="a33183af-95fe-474c-ae5d-d288ab19c967" containerID="3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d" exitCode=0 Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.844327 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twxcz" event={"ID":"a33183af-95fe-474c-ae5d-d288ab19c967","Type":"ContainerDied","Data":"3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d"} Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.844372 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twxcz" event={"ID":"a33183af-95fe-474c-ae5d-d288ab19c967","Type":"ContainerDied","Data":"6568aa07719027300c96660ec1142f945c49928bc44f0bfe3daf23c5ee00d613"} Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.844395 4935 scope.go:117] "RemoveContainer" containerID="3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.844410 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twxcz" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.892962 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twxcz"] Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.900190 4935 scope.go:117] "RemoveContainer" containerID="01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.901920 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-twxcz"] Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.929013 4935 scope.go:117] "RemoveContainer" containerID="7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.969729 4935 scope.go:117] "RemoveContainer" containerID="3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d" Oct 05 09:22:09 crc kubenswrapper[4935]: E1005 09:22:09.970214 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d\": container with ID starting with 3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d not found: ID does not exist" containerID="3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.970256 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d"} err="failed to get container status \"3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d\": rpc error: code = NotFound desc = could not find container \"3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d\": container with ID starting with 3a14858f31c05aa0f6a5a5713752e5e61a9949df0624b65331dffddeb48c560d not found: ID does not exist" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.970301 4935 scope.go:117] "RemoveContainer" containerID="01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50" Oct 05 09:22:09 crc kubenswrapper[4935]: E1005 09:22:09.970786 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50\": container with ID starting with 01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50 not found: ID does not exist" containerID="01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.970815 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50"} err="failed to get container status \"01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50\": rpc error: code = NotFound desc = could not find container \"01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50\": container with ID starting with 01725dbafb12661fbf20a6c833b955373a910e664b75ddd578c9d6a1e8c59c50 not found: ID does not exist" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.970842 4935 scope.go:117] "RemoveContainer" containerID="7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb" Oct 05 09:22:09 crc kubenswrapper[4935]: E1005 09:22:09.971289 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb\": container with ID starting with 7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb not found: ID does not exist" containerID="7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb" Oct 05 09:22:09 crc kubenswrapper[4935]: I1005 09:22:09.971312 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb"} err="failed to get container status \"7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb\": rpc error: code = NotFound desc = could not find container \"7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb\": container with ID starting with 7e2de3b15de9ed230f366a5dda343dcfdb43695b749debe09a5916ce4fb064cb not found: ID does not exist" Oct 05 09:22:10 crc kubenswrapper[4935]: I1005 09:22:10.797542 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" path="/var/lib/kubelet/pods/a33183af-95fe-474c-ae5d-d288ab19c967/volumes" Oct 05 09:22:39 crc kubenswrapper[4935]: I1005 09:22:39.161580 4935 generic.go:334] "Generic (PLEG): container finished" podID="7b5b8e58-6662-4dac-8700-f324b232653b" containerID="0196ac346180e43dcc2a063ca1d829ec45adc7c5f171f61173b8e04d1e1d31a2" exitCode=0 Oct 05 09:22:39 crc kubenswrapper[4935]: I1005 09:22:39.161706 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-zgwf7" event={"ID":"7b5b8e58-6662-4dac-8700-f324b232653b","Type":"ContainerDied","Data":"0196ac346180e43dcc2a063ca1d829ec45adc7c5f171f61173b8e04d1e1d31a2"} Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.061884 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.105649 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ssh-key\") pod \"7b5b8e58-6662-4dac-8700-f324b232653b\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.105803 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ovn-combined-ca-bundle\") pod \"7b5b8e58-6662-4dac-8700-f324b232653b\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.105846 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z4zz\" (UniqueName: \"kubernetes.io/projected/7b5b8e58-6662-4dac-8700-f324b232653b-kube-api-access-2z4zz\") pod \"7b5b8e58-6662-4dac-8700-f324b232653b\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.105952 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-inventory\") pod \"7b5b8e58-6662-4dac-8700-f324b232653b\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.106008 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7b5b8e58-6662-4dac-8700-f324b232653b-ovncontroller-config-0\") pod \"7b5b8e58-6662-4dac-8700-f324b232653b\" (UID: \"7b5b8e58-6662-4dac-8700-f324b232653b\") " Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.111537 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b5b8e58-6662-4dac-8700-f324b232653b-kube-api-access-2z4zz" (OuterVolumeSpecName: "kube-api-access-2z4zz") pod "7b5b8e58-6662-4dac-8700-f324b232653b" (UID: "7b5b8e58-6662-4dac-8700-f324b232653b"). InnerVolumeSpecName "kube-api-access-2z4zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.111553 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7b5b8e58-6662-4dac-8700-f324b232653b" (UID: "7b5b8e58-6662-4dac-8700-f324b232653b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.133553 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b5b8e58-6662-4dac-8700-f324b232653b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "7b5b8e58-6662-4dac-8700-f324b232653b" (UID: "7b5b8e58-6662-4dac-8700-f324b232653b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.137025 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-inventory" (OuterVolumeSpecName: "inventory") pod "7b5b8e58-6662-4dac-8700-f324b232653b" (UID: "7b5b8e58-6662-4dac-8700-f324b232653b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.143849 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b5b8e58-6662-4dac-8700-f324b232653b" (UID: "7b5b8e58-6662-4dac-8700-f324b232653b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.208318 4935 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7b5b8e58-6662-4dac-8700-f324b232653b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.208362 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.208378 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.208390 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z4zz\" (UniqueName: \"kubernetes.io/projected/7b5b8e58-6662-4dac-8700-f324b232653b-kube-api-access-2z4zz\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.208403 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b5b8e58-6662-4dac-8700-f324b232653b-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.212472 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-zgwf7" event={"ID":"7b5b8e58-6662-4dac-8700-f324b232653b","Type":"ContainerDied","Data":"eed14309720934501f514df0e81f23e2efd10b357e361675a37dbb3cc7ed3420"} Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.212514 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eed14309720934501f514df0e81f23e2efd10b357e361675a37dbb3cc7ed3420" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.212548 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-zgwf7" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.323015 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-kr2ff"] Oct 05 09:22:41 crc kubenswrapper[4935]: E1005 09:22:41.323487 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" containerName="extract-utilities" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.323506 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" containerName="extract-utilities" Oct 05 09:22:41 crc kubenswrapper[4935]: E1005 09:22:41.323529 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" containerName="registry-server" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.323536 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" containerName="registry-server" Oct 05 09:22:41 crc kubenswrapper[4935]: E1005 09:22:41.323545 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" containerName="extract-content" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.323552 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" containerName="extract-content" Oct 05 09:22:41 crc kubenswrapper[4935]: E1005 09:22:41.323566 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b5b8e58-6662-4dac-8700-f324b232653b" containerName="ovn-openstack-openstack-networker" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.323573 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b5b8e58-6662-4dac-8700-f324b232653b" containerName="ovn-openstack-openstack-networker" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.323767 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33183af-95fe-474c-ae5d-d288ab19c967" containerName="registry-server" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.323788 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b5b8e58-6662-4dac-8700-f324b232653b" containerName="ovn-openstack-openstack-networker" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.324544 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.326908 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-p45zw" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.327030 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.327081 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.327144 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.334534 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-kr2ff"] Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.421699 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5l2s\" (UniqueName: \"kubernetes.io/projected/e903bbab-dcb0-4fc8-9265-195010d8f46e-kube-api-access-k5l2s\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.421754 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-inventory\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.421785 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.421922 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.422118 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.422318 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.523909 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.523997 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.524060 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.524087 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5l2s\" (UniqueName: \"kubernetes.io/projected/e903bbab-dcb0-4fc8-9265-195010d8f46e-kube-api-access-k5l2s\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.524116 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-inventory\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.524132 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.527962 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-inventory\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.528447 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.529394 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.529554 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.530309 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.548563 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5l2s\" (UniqueName: \"kubernetes.io/projected/e903bbab-dcb0-4fc8-9265-195010d8f46e-kube-api-access-k5l2s\") pod \"neutron-metadata-openstack-openstack-networker-kr2ff\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:41 crc kubenswrapper[4935]: I1005 09:22:41.654131 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:22:42 crc kubenswrapper[4935]: I1005 09:22:42.185305 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-kr2ff"] Oct 05 09:22:43 crc kubenswrapper[4935]: I1005 09:22:43.238949 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" event={"ID":"e903bbab-dcb0-4fc8-9265-195010d8f46e","Type":"ContainerStarted","Data":"9bedc46bbcbe57e57fe229c9f388ac41a069baa9fbb2fc11fd60f0d0ef2754df"} Oct 05 09:22:44 crc kubenswrapper[4935]: I1005 09:22:44.249862 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" event={"ID":"e903bbab-dcb0-4fc8-9265-195010d8f46e","Type":"ContainerStarted","Data":"ab1da4df15a854c4dbc56033d0f6679ce41a38b822693751e1cf650d308384dd"} Oct 05 09:22:44 crc kubenswrapper[4935]: I1005 09:22:44.271729 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" podStartSLOduration=2.833080244 podStartE2EDuration="3.271706222s" podCreationTimestamp="2025-10-05 09:22:41 +0000 UTC" firstStartedPulling="2025-10-05 09:22:42.771456921 +0000 UTC m=+8996.654083421" lastFinishedPulling="2025-10-05 09:22:43.210082939 +0000 UTC m=+8997.092709399" observedRunningTime="2025-10-05 09:22:44.265976141 +0000 UTC m=+8998.148602631" watchObservedRunningTime="2025-10-05 09:22:44.271706222 +0000 UTC m=+8998.154332692" Oct 05 09:22:44 crc kubenswrapper[4935]: I1005 09:22:44.289220 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:22:44 crc kubenswrapper[4935]: I1005 09:22:44.289298 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:23:14 crc kubenswrapper[4935]: I1005 09:23:14.289748 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:23:14 crc kubenswrapper[4935]: I1005 09:23:14.290754 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:23:19 crc kubenswrapper[4935]: I1005 09:23:19.616397 4935 generic.go:334] "Generic (PLEG): container finished" podID="48e06b46-b928-49eb-a8b8-5b4b5d24925b" containerID="e42a079019f667fadb51cf7abc27500a686649c32b6492553826a2c09c16e295" exitCode=0 Oct 05 09:23:19 crc kubenswrapper[4935]: I1005 09:23:19.616489 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" event={"ID":"48e06b46-b928-49eb-a8b8-5b4b5d24925b","Type":"ContainerDied","Data":"e42a079019f667fadb51cf7abc27500a686649c32b6492553826a2c09c16e295"} Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.115834 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.179674 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-inventory\") pod \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.179914 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ssh-key\") pod \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.180000 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovn-combined-ca-bundle\") pod \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.180030 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnttb\" (UniqueName: \"kubernetes.io/projected/48e06b46-b928-49eb-a8b8-5b4b5d24925b-kube-api-access-jnttb\") pod \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.180096 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ceph\") pod \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.180127 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovncontroller-config-0\") pod \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\" (UID: \"48e06b46-b928-49eb-a8b8-5b4b5d24925b\") " Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.186004 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ceph" (OuterVolumeSpecName: "ceph") pod "48e06b46-b928-49eb-a8b8-5b4b5d24925b" (UID: "48e06b46-b928-49eb-a8b8-5b4b5d24925b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.186084 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e06b46-b928-49eb-a8b8-5b4b5d24925b-kube-api-access-jnttb" (OuterVolumeSpecName: "kube-api-access-jnttb") pod "48e06b46-b928-49eb-a8b8-5b4b5d24925b" (UID: "48e06b46-b928-49eb-a8b8-5b4b5d24925b"). InnerVolumeSpecName "kube-api-access-jnttb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.198214 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "48e06b46-b928-49eb-a8b8-5b4b5d24925b" (UID: "48e06b46-b928-49eb-a8b8-5b4b5d24925b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.210166 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48e06b46-b928-49eb-a8b8-5b4b5d24925b" (UID: "48e06b46-b928-49eb-a8b8-5b4b5d24925b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.216699 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-inventory" (OuterVolumeSpecName: "inventory") pod "48e06b46-b928-49eb-a8b8-5b4b5d24925b" (UID: "48e06b46-b928-49eb-a8b8-5b4b5d24925b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.220200 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "48e06b46-b928-49eb-a8b8-5b4b5d24925b" (UID: "48e06b46-b928-49eb-a8b8-5b4b5d24925b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.283331 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.283396 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.283418 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnttb\" (UniqueName: \"kubernetes.io/projected/48e06b46-b928-49eb-a8b8-5b4b5d24925b-kube-api-access-jnttb\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.283437 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.283456 4935 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/48e06b46-b928-49eb-a8b8-5b4b5d24925b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.283473 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e06b46-b928-49eb-a8b8-5b4b5d24925b-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.635400 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" event={"ID":"48e06b46-b928-49eb-a8b8-5b4b5d24925b","Type":"ContainerDied","Data":"6df551988ebc692e574aa0a89867e619148598e06d591451a1a76ff9b8b2db66"} Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.635450 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6df551988ebc692e574aa0a89867e619148598e06d591451a1a76ff9b8b2db66" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.635454 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-lfkb6" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.743499 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-65rhl"] Oct 05 09:23:21 crc kubenswrapper[4935]: E1005 09:23:21.744230 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e06b46-b928-49eb-a8b8-5b4b5d24925b" containerName="ovn-openstack-openstack-cell1" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.744342 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e06b46-b928-49eb-a8b8-5b4b5d24925b" containerName="ovn-openstack-openstack-cell1" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.744605 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e06b46-b928-49eb-a8b8-5b4b5d24925b" containerName="ovn-openstack-openstack-cell1" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.745504 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.747862 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.748144 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.763668 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-65rhl"] Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.790848 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.791688 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.791759 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.791921 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sfcp\" (UniqueName: \"kubernetes.io/projected/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-kube-api-access-9sfcp\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.791954 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.792051 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.792111 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.894274 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.894354 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.894389 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.894471 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.894518 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.894545 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.894618 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sfcp\" (UniqueName: \"kubernetes.io/projected/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-kube-api-access-9sfcp\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.899207 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.899259 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.903370 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.903506 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.903952 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.905964 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:21 crc kubenswrapper[4935]: I1005 09:23:21.912809 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sfcp\" (UniqueName: \"kubernetes.io/projected/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-kube-api-access-9sfcp\") pod \"neutron-metadata-openstack-openstack-cell1-65rhl\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:22 crc kubenswrapper[4935]: I1005 09:23:22.087541 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:23:23 crc kubenswrapper[4935]: I1005 09:23:23.002100 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-65rhl"] Oct 05 09:23:23 crc kubenswrapper[4935]: I1005 09:23:23.658617 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" event={"ID":"c01b53f0-a632-4a5e-bd1c-cae6de7729f3","Type":"ContainerStarted","Data":"1ae4e773b21968309b33222703b6632fd0c8a900ff34e746d64ed9560873cbd6"} Oct 05 09:23:24 crc kubenswrapper[4935]: I1005 09:23:24.671634 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" event={"ID":"c01b53f0-a632-4a5e-bd1c-cae6de7729f3","Type":"ContainerStarted","Data":"c91d40b142fb3363aeec214e599523eaa5718e9fed7b23468fd4f8919849d485"} Oct 05 09:23:24 crc kubenswrapper[4935]: I1005 09:23:24.702805 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" podStartSLOduration=3.228017558 podStartE2EDuration="3.702783274s" podCreationTimestamp="2025-10-05 09:23:21 +0000 UTC" firstStartedPulling="2025-10-05 09:23:23.004248859 +0000 UTC m=+9036.886875319" lastFinishedPulling="2025-10-05 09:23:23.479014575 +0000 UTC m=+9037.361641035" observedRunningTime="2025-10-05 09:23:24.693835207 +0000 UTC m=+9038.576461667" watchObservedRunningTime="2025-10-05 09:23:24.702783274 +0000 UTC m=+9038.585409734" Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.289807 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.290367 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.290413 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.291284 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.291357 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" gracePeriod=600 Oct 05 09:23:44 crc kubenswrapper[4935]: E1005 09:23:44.432244 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.883667 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" exitCode=0 Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.883735 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b"} Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.884032 4935 scope.go:117] "RemoveContainer" containerID="989e4812ce9b8760f616df4a134bbde12de0cc616591fdb55b6764cc85626afe" Oct 05 09:23:44 crc kubenswrapper[4935]: I1005 09:23:44.885245 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:23:44 crc kubenswrapper[4935]: E1005 09:23:44.886105 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:23:46 crc kubenswrapper[4935]: I1005 09:23:46.907359 4935 generic.go:334] "Generic (PLEG): container finished" podID="e903bbab-dcb0-4fc8-9265-195010d8f46e" containerID="ab1da4df15a854c4dbc56033d0f6679ce41a38b822693751e1cf650d308384dd" exitCode=0 Oct 05 09:23:46 crc kubenswrapper[4935]: I1005 09:23:46.907414 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" event={"ID":"e903bbab-dcb0-4fc8-9265-195010d8f46e","Type":"ContainerDied","Data":"ab1da4df15a854c4dbc56033d0f6679ce41a38b822693751e1cf650d308384dd"} Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.442728 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.452430 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"e903bbab-dcb0-4fc8-9265-195010d8f46e\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.452508 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-nova-metadata-neutron-config-0\") pod \"e903bbab-dcb0-4fc8-9265-195010d8f46e\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.452558 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-metadata-combined-ca-bundle\") pod \"e903bbab-dcb0-4fc8-9265-195010d8f46e\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.452588 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5l2s\" (UniqueName: \"kubernetes.io/projected/e903bbab-dcb0-4fc8-9265-195010d8f46e-kube-api-access-k5l2s\") pod \"e903bbab-dcb0-4fc8-9265-195010d8f46e\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.452638 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-inventory\") pod \"e903bbab-dcb0-4fc8-9265-195010d8f46e\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.452880 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-ssh-key\") pod \"e903bbab-dcb0-4fc8-9265-195010d8f46e\" (UID: \"e903bbab-dcb0-4fc8-9265-195010d8f46e\") " Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.475120 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e903bbab-dcb0-4fc8-9265-195010d8f46e-kube-api-access-k5l2s" (OuterVolumeSpecName: "kube-api-access-k5l2s") pod "e903bbab-dcb0-4fc8-9265-195010d8f46e" (UID: "e903bbab-dcb0-4fc8-9265-195010d8f46e"). InnerVolumeSpecName "kube-api-access-k5l2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.475340 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e903bbab-dcb0-4fc8-9265-195010d8f46e" (UID: "e903bbab-dcb0-4fc8-9265-195010d8f46e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.485446 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "e903bbab-dcb0-4fc8-9265-195010d8f46e" (UID: "e903bbab-dcb0-4fc8-9265-195010d8f46e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.485926 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e903bbab-dcb0-4fc8-9265-195010d8f46e" (UID: "e903bbab-dcb0-4fc8-9265-195010d8f46e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.503880 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-inventory" (OuterVolumeSpecName: "inventory") pod "e903bbab-dcb0-4fc8-9265-195010d8f46e" (UID: "e903bbab-dcb0-4fc8-9265-195010d8f46e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.513523 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "e903bbab-dcb0-4fc8-9265-195010d8f46e" (UID: "e903bbab-dcb0-4fc8-9265-195010d8f46e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.556396 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.556553 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.556613 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.556670 4935 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.556723 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e903bbab-dcb0-4fc8-9265-195010d8f46e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.556830 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5l2s\" (UniqueName: \"kubernetes.io/projected/e903bbab-dcb0-4fc8-9265-195010d8f46e-kube-api-access-k5l2s\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.940844 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" event={"ID":"e903bbab-dcb0-4fc8-9265-195010d8f46e","Type":"ContainerDied","Data":"9bedc46bbcbe57e57fe229c9f388ac41a069baa9fbb2fc11fd60f0d0ef2754df"} Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.940929 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bedc46bbcbe57e57fe229c9f388ac41a069baa9fbb2fc11fd60f0d0ef2754df" Oct 05 09:23:48 crc kubenswrapper[4935]: I1005 09:23:48.940954 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-kr2ff" Oct 05 09:23:56 crc kubenswrapper[4935]: I1005 09:23:56.783511 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:23:56 crc kubenswrapper[4935]: E1005 09:23:56.784265 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:24:11 crc kubenswrapper[4935]: I1005 09:24:11.777751 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:24:11 crc kubenswrapper[4935]: E1005 09:24:11.778553 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:24:20 crc kubenswrapper[4935]: I1005 09:24:20.271171 4935 generic.go:334] "Generic (PLEG): container finished" podID="c01b53f0-a632-4a5e-bd1c-cae6de7729f3" containerID="c91d40b142fb3363aeec214e599523eaa5718e9fed7b23468fd4f8919849d485" exitCode=0 Oct 05 09:24:20 crc kubenswrapper[4935]: I1005 09:24:20.271381 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" event={"ID":"c01b53f0-a632-4a5e-bd1c-cae6de7729f3","Type":"ContainerDied","Data":"c91d40b142fb3363aeec214e599523eaa5718e9fed7b23468fd4f8919849d485"} Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.855477 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.964703 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sfcp\" (UniqueName: \"kubernetes.io/projected/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-kube-api-access-9sfcp\") pod \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.964817 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-metadata-combined-ca-bundle\") pod \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.964851 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-inventory\") pod \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.964924 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-nova-metadata-neutron-config-0\") pod \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.965035 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ssh-key\") pod \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.965112 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ceph\") pod \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.965212 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\" (UID: \"c01b53f0-a632-4a5e-bd1c-cae6de7729f3\") " Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.969832 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-kube-api-access-9sfcp" (OuterVolumeSpecName: "kube-api-access-9sfcp") pod "c01b53f0-a632-4a5e-bd1c-cae6de7729f3" (UID: "c01b53f0-a632-4a5e-bd1c-cae6de7729f3"). InnerVolumeSpecName "kube-api-access-9sfcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.970061 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c01b53f0-a632-4a5e-bd1c-cae6de7729f3" (UID: "c01b53f0-a632-4a5e-bd1c-cae6de7729f3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.970346 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ceph" (OuterVolumeSpecName: "ceph") pod "c01b53f0-a632-4a5e-bd1c-cae6de7729f3" (UID: "c01b53f0-a632-4a5e-bd1c-cae6de7729f3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:24:21 crc kubenswrapper[4935]: I1005 09:24:21.993130 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-inventory" (OuterVolumeSpecName: "inventory") pod "c01b53f0-a632-4a5e-bd1c-cae6de7729f3" (UID: "c01b53f0-a632-4a5e-bd1c-cae6de7729f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.000836 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "c01b53f0-a632-4a5e-bd1c-cae6de7729f3" (UID: "c01b53f0-a632-4a5e-bd1c-cae6de7729f3"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.001880 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "c01b53f0-a632-4a5e-bd1c-cae6de7729f3" (UID: "c01b53f0-a632-4a5e-bd1c-cae6de7729f3"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.003385 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c01b53f0-a632-4a5e-bd1c-cae6de7729f3" (UID: "c01b53f0-a632-4a5e-bd1c-cae6de7729f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.070257 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.070308 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.070454 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sfcp\" (UniqueName: \"kubernetes.io/projected/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-kube-api-access-9sfcp\") on node \"crc\" DevicePath \"\"" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.070627 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.070654 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.070666 4935 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.070678 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c01b53f0-a632-4a5e-bd1c-cae6de7729f3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.321982 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" event={"ID":"c01b53f0-a632-4a5e-bd1c-cae6de7729f3","Type":"ContainerDied","Data":"1ae4e773b21968309b33222703b6632fd0c8a900ff34e746d64ed9560873cbd6"} Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.322043 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ae4e773b21968309b33222703b6632fd0c8a900ff34e746d64ed9560873cbd6" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.322124 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-65rhl" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.445864 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-fslpj"] Oct 05 09:24:22 crc kubenswrapper[4935]: E1005 09:24:22.446745 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e903bbab-dcb0-4fc8-9265-195010d8f46e" containerName="neutron-metadata-openstack-openstack-networker" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.446922 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e903bbab-dcb0-4fc8-9265-195010d8f46e" containerName="neutron-metadata-openstack-openstack-networker" Oct 05 09:24:22 crc kubenswrapper[4935]: E1005 09:24:22.447109 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b53f0-a632-4a5e-bd1c-cae6de7729f3" containerName="neutron-metadata-openstack-openstack-cell1" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.447221 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b53f0-a632-4a5e-bd1c-cae6de7729f3" containerName="neutron-metadata-openstack-openstack-cell1" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.447724 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b53f0-a632-4a5e-bd1c-cae6de7729f3" containerName="neutron-metadata-openstack-openstack-cell1" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.447860 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e903bbab-dcb0-4fc8-9265-195010d8f46e" containerName="neutron-metadata-openstack-openstack-networker" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.449334 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.452481 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.458253 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-fslpj"] Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.452656 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.452714 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.452762 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.456794 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.581583 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ssh-key\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.581958 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-inventory\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.582336 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ceph\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.582441 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.582544 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl4dn\" (UniqueName: \"kubernetes.io/projected/4eec0ed5-4af6-45bd-9196-e3ebd559575b-kube-api-access-vl4dn\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.582692 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.685178 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ceph\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.685471 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.685519 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl4dn\" (UniqueName: \"kubernetes.io/projected/4eec0ed5-4af6-45bd-9196-e3ebd559575b-kube-api-access-vl4dn\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.685571 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.685676 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ssh-key\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.685847 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-inventory\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.688962 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ssh-key\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.689374 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ceph\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.689389 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.689420 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-inventory\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.690931 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.703373 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl4dn\" (UniqueName: \"kubernetes.io/projected/4eec0ed5-4af6-45bd-9196-e3ebd559575b-kube-api-access-vl4dn\") pod \"libvirt-openstack-openstack-cell1-fslpj\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:22 crc kubenswrapper[4935]: I1005 09:24:22.769425 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:24:23 crc kubenswrapper[4935]: I1005 09:24:23.395061 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-fslpj"] Oct 05 09:24:23 crc kubenswrapper[4935]: W1005 09:24:23.399472 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eec0ed5_4af6_45bd_9196_e3ebd559575b.slice/crio-5bf275e2a1438a8ce350ce4c21d4010597d23c7f2ec2e348d12312519f17444b WatchSource:0}: Error finding container 5bf275e2a1438a8ce350ce4c21d4010597d23c7f2ec2e348d12312519f17444b: Status 404 returned error can't find the container with id 5bf275e2a1438a8ce350ce4c21d4010597d23c7f2ec2e348d12312519f17444b Oct 05 09:24:24 crc kubenswrapper[4935]: I1005 09:24:24.352166 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" event={"ID":"4eec0ed5-4af6-45bd-9196-e3ebd559575b","Type":"ContainerStarted","Data":"ffaebc6117bf9edc2e93dec9d7a5f3646cf9d6b51ce7a3960ad77925d761e490"} Oct 05 09:24:24 crc kubenswrapper[4935]: I1005 09:24:24.352736 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" event={"ID":"4eec0ed5-4af6-45bd-9196-e3ebd559575b","Type":"ContainerStarted","Data":"5bf275e2a1438a8ce350ce4c21d4010597d23c7f2ec2e348d12312519f17444b"} Oct 05 09:24:24 crc kubenswrapper[4935]: I1005 09:24:24.377643 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" podStartSLOduration=1.910886177 podStartE2EDuration="2.377626579s" podCreationTimestamp="2025-10-05 09:24:22 +0000 UTC" firstStartedPulling="2025-10-05 09:24:23.409523946 +0000 UTC m=+9097.292150406" lastFinishedPulling="2025-10-05 09:24:23.876264348 +0000 UTC m=+9097.758890808" observedRunningTime="2025-10-05 09:24:24.370228723 +0000 UTC m=+9098.252855183" watchObservedRunningTime="2025-10-05 09:24:24.377626579 +0000 UTC m=+9098.260253039" Oct 05 09:24:25 crc kubenswrapper[4935]: I1005 09:24:25.777884 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:24:25 crc kubenswrapper[4935]: E1005 09:24:25.778714 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:24:38 crc kubenswrapper[4935]: I1005 09:24:38.777545 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:24:38 crc kubenswrapper[4935]: E1005 09:24:38.778629 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:24:49 crc kubenswrapper[4935]: I1005 09:24:49.788316 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:24:49 crc kubenswrapper[4935]: E1005 09:24:49.790327 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:25:02 crc kubenswrapper[4935]: I1005 09:25:02.778096 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:25:02 crc kubenswrapper[4935]: E1005 09:25:02.779313 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:25:17 crc kubenswrapper[4935]: I1005 09:25:17.778321 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:25:17 crc kubenswrapper[4935]: E1005 09:25:17.779186 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:25:28 crc kubenswrapper[4935]: I1005 09:25:28.778141 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:25:28 crc kubenswrapper[4935]: E1005 09:25:28.779194 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:25:40 crc kubenswrapper[4935]: I1005 09:25:40.777605 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:25:40 crc kubenswrapper[4935]: E1005 09:25:40.778299 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:25:55 crc kubenswrapper[4935]: I1005 09:25:55.776922 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:25:55 crc kubenswrapper[4935]: E1005 09:25:55.777735 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:26:05 crc kubenswrapper[4935]: I1005 09:26:05.844523 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jftck"] Oct 05 09:26:05 crc kubenswrapper[4935]: I1005 09:26:05.852288 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:05 crc kubenswrapper[4935]: I1005 09:26:05.893850 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jftck"] Oct 05 09:26:05 crc kubenswrapper[4935]: I1005 09:26:05.922204 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-catalog-content\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:05 crc kubenswrapper[4935]: I1005 09:26:05.923247 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-utilities\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:05 crc kubenswrapper[4935]: I1005 09:26:05.923449 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qml7d\" (UniqueName: \"kubernetes.io/projected/f6770c90-2229-48e0-8e7a-421da15118a3-kube-api-access-qml7d\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.025920 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-utilities\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.026032 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qml7d\" (UniqueName: \"kubernetes.io/projected/f6770c90-2229-48e0-8e7a-421da15118a3-kube-api-access-qml7d\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.026098 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-catalog-content\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.026591 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-utilities\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.026637 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-catalog-content\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.046800 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qml7d\" (UniqueName: \"kubernetes.io/projected/f6770c90-2229-48e0-8e7a-421da15118a3-kube-api-access-qml7d\") pod \"redhat-operators-jftck\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.234755 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.756987 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jftck"] Oct 05 09:26:06 crc kubenswrapper[4935]: I1005 09:26:06.790187 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:26:06 crc kubenswrapper[4935]: E1005 09:26:06.790436 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:26:07 crc kubenswrapper[4935]: I1005 09:26:07.567411 4935 generic.go:334] "Generic (PLEG): container finished" podID="f6770c90-2229-48e0-8e7a-421da15118a3" containerID="5bfb8ff2b2c76654cfe4f06053e2f97368c99078f0e73f0387753500632d639f" exitCode=0 Oct 05 09:26:07 crc kubenswrapper[4935]: I1005 09:26:07.567928 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jftck" event={"ID":"f6770c90-2229-48e0-8e7a-421da15118a3","Type":"ContainerDied","Data":"5bfb8ff2b2c76654cfe4f06053e2f97368c99078f0e73f0387753500632d639f"} Oct 05 09:26:07 crc kubenswrapper[4935]: I1005 09:26:07.567956 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jftck" event={"ID":"f6770c90-2229-48e0-8e7a-421da15118a3","Type":"ContainerStarted","Data":"de5d8a657a6ba43d59a4d7066a14ca23b22d26cd9acf3946c48366d2c1102041"} Oct 05 09:26:07 crc kubenswrapper[4935]: I1005 09:26:07.570433 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:26:08 crc kubenswrapper[4935]: I1005 09:26:08.578486 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jftck" event={"ID":"f6770c90-2229-48e0-8e7a-421da15118a3","Type":"ContainerStarted","Data":"3e89884e9368281e78a4197d702229690a14b051785d183918dd5ad55bc9ed51"} Oct 05 09:26:12 crc kubenswrapper[4935]: I1005 09:26:12.617591 4935 generic.go:334] "Generic (PLEG): container finished" podID="f6770c90-2229-48e0-8e7a-421da15118a3" containerID="3e89884e9368281e78a4197d702229690a14b051785d183918dd5ad55bc9ed51" exitCode=0 Oct 05 09:26:12 crc kubenswrapper[4935]: I1005 09:26:12.617686 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jftck" event={"ID":"f6770c90-2229-48e0-8e7a-421da15118a3","Type":"ContainerDied","Data":"3e89884e9368281e78a4197d702229690a14b051785d183918dd5ad55bc9ed51"} Oct 05 09:26:13 crc kubenswrapper[4935]: I1005 09:26:13.629470 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jftck" event={"ID":"f6770c90-2229-48e0-8e7a-421da15118a3","Type":"ContainerStarted","Data":"a54253299ced6e2085abcf3fdd16e87a5937fdd339872ba042d8c2d91cb8ee5a"} Oct 05 09:26:13 crc kubenswrapper[4935]: I1005 09:26:13.648066 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jftck" podStartSLOduration=3.149088436 podStartE2EDuration="8.648045885s" podCreationTimestamp="2025-10-05 09:26:05 +0000 UTC" firstStartedPulling="2025-10-05 09:26:07.57021262 +0000 UTC m=+9201.452839080" lastFinishedPulling="2025-10-05 09:26:13.069170059 +0000 UTC m=+9206.951796529" observedRunningTime="2025-10-05 09:26:13.646479373 +0000 UTC m=+9207.529105843" watchObservedRunningTime="2025-10-05 09:26:13.648045885 +0000 UTC m=+9207.530672345" Oct 05 09:26:16 crc kubenswrapper[4935]: I1005 09:26:16.236395 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:16 crc kubenswrapper[4935]: I1005 09:26:16.236863 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:17 crc kubenswrapper[4935]: I1005 09:26:17.373837 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jftck" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="registry-server" probeResult="failure" output=< Oct 05 09:26:17 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 09:26:17 crc kubenswrapper[4935]: > Oct 05 09:26:21 crc kubenswrapper[4935]: I1005 09:26:21.776819 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:26:21 crc kubenswrapper[4935]: E1005 09:26:21.778600 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:26:26 crc kubenswrapper[4935]: I1005 09:26:26.610092 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:26 crc kubenswrapper[4935]: I1005 09:26:26.669537 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:26 crc kubenswrapper[4935]: I1005 09:26:26.854954 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jftck"] Oct 05 09:26:27 crc kubenswrapper[4935]: I1005 09:26:27.821463 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jftck" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="registry-server" containerID="cri-o://a54253299ced6e2085abcf3fdd16e87a5937fdd339872ba042d8c2d91cb8ee5a" gracePeriod=2 Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.843136 4935 generic.go:334] "Generic (PLEG): container finished" podID="f6770c90-2229-48e0-8e7a-421da15118a3" containerID="a54253299ced6e2085abcf3fdd16e87a5937fdd339872ba042d8c2d91cb8ee5a" exitCode=0 Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.843232 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jftck" event={"ID":"f6770c90-2229-48e0-8e7a-421da15118a3","Type":"ContainerDied","Data":"a54253299ced6e2085abcf3fdd16e87a5937fdd339872ba042d8c2d91cb8ee5a"} Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.843660 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jftck" event={"ID":"f6770c90-2229-48e0-8e7a-421da15118a3","Type":"ContainerDied","Data":"de5d8a657a6ba43d59a4d7066a14ca23b22d26cd9acf3946c48366d2c1102041"} Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.843678 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de5d8a657a6ba43d59a4d7066a14ca23b22d26cd9acf3946c48366d2c1102041" Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.886260 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.943996 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qml7d\" (UniqueName: \"kubernetes.io/projected/f6770c90-2229-48e0-8e7a-421da15118a3-kube-api-access-qml7d\") pod \"f6770c90-2229-48e0-8e7a-421da15118a3\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.944227 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-catalog-content\") pod \"f6770c90-2229-48e0-8e7a-421da15118a3\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.944356 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-utilities\") pod \"f6770c90-2229-48e0-8e7a-421da15118a3\" (UID: \"f6770c90-2229-48e0-8e7a-421da15118a3\") " Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.945229 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-utilities" (OuterVolumeSpecName: "utilities") pod "f6770c90-2229-48e0-8e7a-421da15118a3" (UID: "f6770c90-2229-48e0-8e7a-421da15118a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:26:28 crc kubenswrapper[4935]: I1005 09:26:28.950097 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6770c90-2229-48e0-8e7a-421da15118a3-kube-api-access-qml7d" (OuterVolumeSpecName: "kube-api-access-qml7d") pod "f6770c90-2229-48e0-8e7a-421da15118a3" (UID: "f6770c90-2229-48e0-8e7a-421da15118a3"). InnerVolumeSpecName "kube-api-access-qml7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:26:29 crc kubenswrapper[4935]: I1005 09:26:29.043217 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6770c90-2229-48e0-8e7a-421da15118a3" (UID: "f6770c90-2229-48e0-8e7a-421da15118a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:26:29 crc kubenswrapper[4935]: I1005 09:26:29.047130 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qml7d\" (UniqueName: \"kubernetes.io/projected/f6770c90-2229-48e0-8e7a-421da15118a3-kube-api-access-qml7d\") on node \"crc\" DevicePath \"\"" Oct 05 09:26:29 crc kubenswrapper[4935]: I1005 09:26:29.047185 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:26:29 crc kubenswrapper[4935]: I1005 09:26:29.047202 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6770c90-2229-48e0-8e7a-421da15118a3-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:26:29 crc kubenswrapper[4935]: I1005 09:26:29.855113 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jftck" Oct 05 09:26:29 crc kubenswrapper[4935]: I1005 09:26:29.894357 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jftck"] Oct 05 09:26:29 crc kubenswrapper[4935]: I1005 09:26:29.907176 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jftck"] Oct 05 09:26:30 crc kubenswrapper[4935]: I1005 09:26:30.796105 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" path="/var/lib/kubelet/pods/f6770c90-2229-48e0-8e7a-421da15118a3/volumes" Oct 05 09:26:35 crc kubenswrapper[4935]: I1005 09:26:35.777806 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:26:35 crc kubenswrapper[4935]: E1005 09:26:35.778566 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:26:46 crc kubenswrapper[4935]: I1005 09:26:46.796939 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:26:46 crc kubenswrapper[4935]: E1005 09:26:46.798046 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:26:57 crc kubenswrapper[4935]: I1005 09:26:57.776680 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:26:57 crc kubenswrapper[4935]: E1005 09:26:57.777391 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:27:11 crc kubenswrapper[4935]: I1005 09:27:11.776983 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:27:11 crc kubenswrapper[4935]: E1005 09:27:11.777594 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:27:26 crc kubenswrapper[4935]: I1005 09:27:26.787622 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:27:26 crc kubenswrapper[4935]: E1005 09:27:26.788375 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:27:41 crc kubenswrapper[4935]: I1005 09:27:41.776606 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:27:41 crc kubenswrapper[4935]: E1005 09:27:41.777364 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:27:55 crc kubenswrapper[4935]: I1005 09:27:55.778711 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:27:55 crc kubenswrapper[4935]: E1005 09:27:55.780573 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:28:06 crc kubenswrapper[4935]: I1005 09:28:06.788610 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:28:06 crc kubenswrapper[4935]: E1005 09:28:06.789543 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.501699 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j8hh4"] Oct 05 09:28:11 crc kubenswrapper[4935]: E1005 09:28:11.502850 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="extract-utilities" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.502871 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="extract-utilities" Oct 05 09:28:11 crc kubenswrapper[4935]: E1005 09:28:11.502906 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="registry-server" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.502915 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="registry-server" Oct 05 09:28:11 crc kubenswrapper[4935]: E1005 09:28:11.502943 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="extract-content" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.502951 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="extract-content" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.503208 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6770c90-2229-48e0-8e7a-421da15118a3" containerName="registry-server" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.505419 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.522234 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j8hh4"] Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.637684 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-catalog-content\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.637747 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-utilities\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.638082 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjhs5\" (UniqueName: \"kubernetes.io/projected/6b812fb9-3483-4928-bd94-00a5c7f43c76-kube-api-access-qjhs5\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.740187 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-catalog-content\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.740266 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-utilities\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.740368 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjhs5\" (UniqueName: \"kubernetes.io/projected/6b812fb9-3483-4928-bd94-00a5c7f43c76-kube-api-access-qjhs5\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.741294 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-catalog-content\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:11 crc kubenswrapper[4935]: I1005 09:28:11.741758 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-utilities\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:12 crc kubenswrapper[4935]: I1005 09:28:12.162940 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjhs5\" (UniqueName: \"kubernetes.io/projected/6b812fb9-3483-4928-bd94-00a5c7f43c76-kube-api-access-qjhs5\") pod \"certified-operators-j8hh4\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:12 crc kubenswrapper[4935]: I1005 09:28:12.175388 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:12 crc kubenswrapper[4935]: I1005 09:28:12.699863 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j8hh4"] Oct 05 09:28:13 crc kubenswrapper[4935]: I1005 09:28:13.086127 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerID="71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00" exitCode=0 Oct 05 09:28:13 crc kubenswrapper[4935]: I1005 09:28:13.086169 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8hh4" event={"ID":"6b812fb9-3483-4928-bd94-00a5c7f43c76","Type":"ContainerDied","Data":"71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00"} Oct 05 09:28:13 crc kubenswrapper[4935]: I1005 09:28:13.086440 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8hh4" event={"ID":"6b812fb9-3483-4928-bd94-00a5c7f43c76","Type":"ContainerStarted","Data":"b7bd7070b82f2f3b48e3f2276efaf2564a977c2785d16e37cf55f9ae6091aba5"} Oct 05 09:28:13 crc kubenswrapper[4935]: E1005 09:28:13.098812 4935 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b812fb9_3483_4928_bd94_00a5c7f43c76.slice/crio-conmon-71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00.scope\": RecentStats: unable to find data in memory cache]" Oct 05 09:28:15 crc kubenswrapper[4935]: I1005 09:28:15.119468 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8hh4" event={"ID":"6b812fb9-3483-4928-bd94-00a5c7f43c76","Type":"ContainerStarted","Data":"9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5"} Oct 05 09:28:16 crc kubenswrapper[4935]: I1005 09:28:16.138205 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerID="9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5" exitCode=0 Oct 05 09:28:16 crc kubenswrapper[4935]: I1005 09:28:16.138282 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8hh4" event={"ID":"6b812fb9-3483-4928-bd94-00a5c7f43c76","Type":"ContainerDied","Data":"9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5"} Oct 05 09:28:17 crc kubenswrapper[4935]: I1005 09:28:17.169406 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8hh4" event={"ID":"6b812fb9-3483-4928-bd94-00a5c7f43c76","Type":"ContainerStarted","Data":"51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee"} Oct 05 09:28:17 crc kubenswrapper[4935]: I1005 09:28:17.203733 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j8hh4" podStartSLOduration=2.657328474 podStartE2EDuration="6.203709913s" podCreationTimestamp="2025-10-05 09:28:11 +0000 UTC" firstStartedPulling="2025-10-05 09:28:13.087579401 +0000 UTC m=+9326.970205861" lastFinishedPulling="2025-10-05 09:28:16.63396084 +0000 UTC m=+9330.516587300" observedRunningTime="2025-10-05 09:28:17.1983112 +0000 UTC m=+9331.080937700" watchObservedRunningTime="2025-10-05 09:28:17.203709913 +0000 UTC m=+9331.086336383" Oct 05 09:28:19 crc kubenswrapper[4935]: I1005 09:28:19.778793 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:28:19 crc kubenswrapper[4935]: E1005 09:28:19.780197 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:28:22 crc kubenswrapper[4935]: I1005 09:28:22.176467 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:22 crc kubenswrapper[4935]: I1005 09:28:22.177059 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:22 crc kubenswrapper[4935]: I1005 09:28:22.246522 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:22 crc kubenswrapper[4935]: I1005 09:28:22.319178 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:22 crc kubenswrapper[4935]: I1005 09:28:22.481471 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j8hh4"] Oct 05 09:28:24 crc kubenswrapper[4935]: I1005 09:28:24.244808 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j8hh4" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerName="registry-server" containerID="cri-o://51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee" gracePeriod=2 Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.144582 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.250103 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-catalog-content\") pod \"6b812fb9-3483-4928-bd94-00a5c7f43c76\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.250344 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjhs5\" (UniqueName: \"kubernetes.io/projected/6b812fb9-3483-4928-bd94-00a5c7f43c76-kube-api-access-qjhs5\") pod \"6b812fb9-3483-4928-bd94-00a5c7f43c76\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.250449 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-utilities\") pod \"6b812fb9-3483-4928-bd94-00a5c7f43c76\" (UID: \"6b812fb9-3483-4928-bd94-00a5c7f43c76\") " Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.251479 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-utilities" (OuterVolumeSpecName: "utilities") pod "6b812fb9-3483-4928-bd94-00a5c7f43c76" (UID: "6b812fb9-3483-4928-bd94-00a5c7f43c76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.264208 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b812fb9-3483-4928-bd94-00a5c7f43c76-kube-api-access-qjhs5" (OuterVolumeSpecName: "kube-api-access-qjhs5") pod "6b812fb9-3483-4928-bd94-00a5c7f43c76" (UID: "6b812fb9-3483-4928-bd94-00a5c7f43c76"). InnerVolumeSpecName "kube-api-access-qjhs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.265709 4935 generic.go:334] "Generic (PLEG): container finished" podID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerID="51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee" exitCode=0 Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.265754 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8hh4" event={"ID":"6b812fb9-3483-4928-bd94-00a5c7f43c76","Type":"ContainerDied","Data":"51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee"} Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.265787 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8hh4" event={"ID":"6b812fb9-3483-4928-bd94-00a5c7f43c76","Type":"ContainerDied","Data":"b7bd7070b82f2f3b48e3f2276efaf2564a977c2785d16e37cf55f9ae6091aba5"} Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.265792 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j8hh4" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.265807 4935 scope.go:117] "RemoveContainer" containerID="51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.326111 4935 scope.go:117] "RemoveContainer" containerID="9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.350435 4935 scope.go:117] "RemoveContainer" containerID="71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.355017 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjhs5\" (UniqueName: \"kubernetes.io/projected/6b812fb9-3483-4928-bd94-00a5c7f43c76-kube-api-access-qjhs5\") on node \"crc\" DevicePath \"\"" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.355068 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b812fb9-3483-4928-bd94-00a5c7f43c76" (UID: "6b812fb9-3483-4928-bd94-00a5c7f43c76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.356155 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.395765 4935 scope.go:117] "RemoveContainer" containerID="51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee" Oct 05 09:28:25 crc kubenswrapper[4935]: E1005 09:28:25.396300 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee\": container with ID starting with 51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee not found: ID does not exist" containerID="51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.396356 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee"} err="failed to get container status \"51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee\": rpc error: code = NotFound desc = could not find container \"51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee\": container with ID starting with 51c310f991336185467e55c50ecbfc6e50d25e57300e612891e8725134d094ee not found: ID does not exist" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.396398 4935 scope.go:117] "RemoveContainer" containerID="9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5" Oct 05 09:28:25 crc kubenswrapper[4935]: E1005 09:28:25.396763 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5\": container with ID starting with 9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5 not found: ID does not exist" containerID="9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.396810 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5"} err="failed to get container status \"9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5\": rpc error: code = NotFound desc = could not find container \"9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5\": container with ID starting with 9b1a9b1a64e0f1d5172ad4120cd4cabe0b6a473e5d309ed511385a96684216b5 not found: ID does not exist" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.396838 4935 scope.go:117] "RemoveContainer" containerID="71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00" Oct 05 09:28:25 crc kubenswrapper[4935]: E1005 09:28:25.397150 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00\": container with ID starting with 71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00 not found: ID does not exist" containerID="71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.397182 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00"} err="failed to get container status \"71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00\": rpc error: code = NotFound desc = could not find container \"71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00\": container with ID starting with 71080c05628120cbf4d8aaef43331a6e42c801501baae89e6f1432ae53ba3c00 not found: ID does not exist" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.458416 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b812fb9-3483-4928-bd94-00a5c7f43c76-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.607258 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j8hh4"] Oct 05 09:28:25 crc kubenswrapper[4935]: I1005 09:28:25.618018 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j8hh4"] Oct 05 09:28:26 crc kubenswrapper[4935]: I1005 09:28:26.801550 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" path="/var/lib/kubelet/pods/6b812fb9-3483-4928-bd94-00a5c7f43c76/volumes" Oct 05 09:28:31 crc kubenswrapper[4935]: I1005 09:28:31.777473 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:28:31 crc kubenswrapper[4935]: E1005 09:28:31.778506 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:28:44 crc kubenswrapper[4935]: I1005 09:28:44.777659 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:28:45 crc kubenswrapper[4935]: I1005 09:28:45.512627 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"379bfbe0de420a386519e50fa2c6588e131783fe5dfe591ec5759067e2e26bbc"} Oct 05 09:29:19 crc kubenswrapper[4935]: I1005 09:29:19.922441 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" event={"ID":"4eec0ed5-4af6-45bd-9196-e3ebd559575b","Type":"ContainerDied","Data":"ffaebc6117bf9edc2e93dec9d7a5f3646cf9d6b51ce7a3960ad77925d761e490"} Oct 05 09:29:19 crc kubenswrapper[4935]: I1005 09:29:19.922384 4935 generic.go:334] "Generic (PLEG): container finished" podID="4eec0ed5-4af6-45bd-9196-e3ebd559575b" containerID="ffaebc6117bf9edc2e93dec9d7a5f3646cf9d6b51ce7a3960ad77925d761e490" exitCode=0 Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.614732 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.695805 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl4dn\" (UniqueName: \"kubernetes.io/projected/4eec0ed5-4af6-45bd-9196-e3ebd559575b-kube-api-access-vl4dn\") pod \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.696011 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-secret-0\") pod \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.696068 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ceph\") pod \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.696106 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ssh-key\") pod \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.696231 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-combined-ca-bundle\") pod \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.696305 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-inventory\") pod \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\" (UID: \"4eec0ed5-4af6-45bd-9196-e3ebd559575b\") " Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.713256 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ceph" (OuterVolumeSpecName: "ceph") pod "4eec0ed5-4af6-45bd-9196-e3ebd559575b" (UID: "4eec0ed5-4af6-45bd-9196-e3ebd559575b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.713331 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eec0ed5-4af6-45bd-9196-e3ebd559575b-kube-api-access-vl4dn" (OuterVolumeSpecName: "kube-api-access-vl4dn") pod "4eec0ed5-4af6-45bd-9196-e3ebd559575b" (UID: "4eec0ed5-4af6-45bd-9196-e3ebd559575b"). InnerVolumeSpecName "kube-api-access-vl4dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.715076 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4eec0ed5-4af6-45bd-9196-e3ebd559575b" (UID: "4eec0ed5-4af6-45bd-9196-e3ebd559575b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.727219 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4eec0ed5-4af6-45bd-9196-e3ebd559575b" (UID: "4eec0ed5-4af6-45bd-9196-e3ebd559575b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.727934 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4eec0ed5-4af6-45bd-9196-e3ebd559575b" (UID: "4eec0ed5-4af6-45bd-9196-e3ebd559575b"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.738097 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-inventory" (OuterVolumeSpecName: "inventory") pod "4eec0ed5-4af6-45bd-9196-e3ebd559575b" (UID: "4eec0ed5-4af6-45bd-9196-e3ebd559575b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.798043 4935 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.798072 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.798082 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.798094 4935 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.798106 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eec0ed5-4af6-45bd-9196-e3ebd559575b-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.798140 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl4dn\" (UniqueName: \"kubernetes.io/projected/4eec0ed5-4af6-45bd-9196-e3ebd559575b-kube-api-access-vl4dn\") on node \"crc\" DevicePath \"\"" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.950420 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" event={"ID":"4eec0ed5-4af6-45bd-9196-e3ebd559575b","Type":"ContainerDied","Data":"5bf275e2a1438a8ce350ce4c21d4010597d23c7f2ec2e348d12312519f17444b"} Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.950474 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-fslpj" Oct 05 09:29:21 crc kubenswrapper[4935]: I1005 09:29:21.950483 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bf275e2a1438a8ce350ce4c21d4010597d23c7f2ec2e348d12312519f17444b" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.243941 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-mrgrf"] Oct 05 09:29:22 crc kubenswrapper[4935]: E1005 09:29:22.244461 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerName="registry-server" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.244480 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerName="registry-server" Oct 05 09:29:22 crc kubenswrapper[4935]: E1005 09:29:22.244513 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerName="extract-content" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.244520 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerName="extract-content" Oct 05 09:29:22 crc kubenswrapper[4935]: E1005 09:29:22.244536 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eec0ed5-4af6-45bd-9196-e3ebd559575b" containerName="libvirt-openstack-openstack-cell1" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.244544 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eec0ed5-4af6-45bd-9196-e3ebd559575b" containerName="libvirt-openstack-openstack-cell1" Oct 05 09:29:22 crc kubenswrapper[4935]: E1005 09:29:22.244560 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerName="extract-utilities" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.244568 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerName="extract-utilities" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.244819 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eec0ed5-4af6-45bd-9196-e3ebd559575b" containerName="libvirt-openstack-openstack-cell1" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.244843 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b812fb9-3483-4928-bd94-00a5c7f43c76" containerName="registry-server" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.245772 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.248935 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.249261 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.249462 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.249697 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.250127 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.261222 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.269025 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-mrgrf"] Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.308471 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326306 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326352 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ceph\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326381 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326403 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-inventory\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326438 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326486 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swbrl\" (UniqueName: \"kubernetes.io/projected/2bad5788-676a-436c-bfad-20d810bbd73e-kube-api-access-swbrl\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326530 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326621 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326719 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326773 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.326809 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.429718 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.429792 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.429845 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.429864 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ceph\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.429888 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.429919 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-inventory\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.429943 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.430135 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swbrl\" (UniqueName: \"kubernetes.io/projected/2bad5788-676a-436c-bfad-20d810bbd73e-kube-api-access-swbrl\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.430170 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.430244 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.430307 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.432393 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.443809 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.863558 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.863651 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-inventory\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.863705 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.863742 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.864618 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.864765 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.864979 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swbrl\" (UniqueName: \"kubernetes.io/projected/2bad5788-676a-436c-bfad-20d810bbd73e-kube-api-access-swbrl\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.866485 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ceph\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.868609 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-mrgrf\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:22 crc kubenswrapper[4935]: I1005 09:29:22.928670 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:29:23 crc kubenswrapper[4935]: I1005 09:29:23.318187 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-mrgrf"] Oct 05 09:29:23 crc kubenswrapper[4935]: I1005 09:29:23.979476 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" event={"ID":"2bad5788-676a-436c-bfad-20d810bbd73e","Type":"ContainerStarted","Data":"011487f79b8aef2161fc1c17f8eba732d33146f5f881e41c9d7d6d5730402920"} Oct 05 09:29:24 crc kubenswrapper[4935]: I1005 09:29:24.993853 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" event={"ID":"2bad5788-676a-436c-bfad-20d810bbd73e","Type":"ContainerStarted","Data":"3c9e98f7c5722d8cbfe652726daa0e19902562e2446d70d866f5069f50f976b1"} Oct 05 09:29:25 crc kubenswrapper[4935]: I1005 09:29:25.020018 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" podStartSLOduration=2.525775928 podStartE2EDuration="3.020000456s" podCreationTimestamp="2025-10-05 09:29:22 +0000 UTC" firstStartedPulling="2025-10-05 09:29:23.328582381 +0000 UTC m=+9397.211208841" lastFinishedPulling="2025-10-05 09:29:23.822806899 +0000 UTC m=+9397.705433369" observedRunningTime="2025-10-05 09:29:25.016168956 +0000 UTC m=+9398.898795426" watchObservedRunningTime="2025-10-05 09:29:25.020000456 +0000 UTC m=+9398.902626916" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.158297 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s"] Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.160815 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.171065 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.172123 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.173112 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s"] Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.268468 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdsll\" (UniqueName: \"kubernetes.io/projected/8d654093-f8b9-4333-8949-52bd35aea8d1-kube-api-access-mdsll\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.269097 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d654093-f8b9-4333-8949-52bd35aea8d1-secret-volume\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.269263 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d654093-f8b9-4333-8949-52bd35aea8d1-config-volume\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.371653 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d654093-f8b9-4333-8949-52bd35aea8d1-secret-volume\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.371728 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d654093-f8b9-4333-8949-52bd35aea8d1-config-volume\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.371878 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdsll\" (UniqueName: \"kubernetes.io/projected/8d654093-f8b9-4333-8949-52bd35aea8d1-kube-api-access-mdsll\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.373484 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d654093-f8b9-4333-8949-52bd35aea8d1-config-volume\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.381658 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d654093-f8b9-4333-8949-52bd35aea8d1-secret-volume\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.390772 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdsll\" (UniqueName: \"kubernetes.io/projected/8d654093-f8b9-4333-8949-52bd35aea8d1-kube-api-access-mdsll\") pod \"collect-profiles-29327610-xnl6s\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.511982 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:00 crc kubenswrapper[4935]: W1005 09:30:00.990643 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d654093_f8b9_4333_8949_52bd35aea8d1.slice/crio-5af196b1dc0d416a7a6ae7e69e119e54ca2dba5a6d17a990a7c40a1d7ced6b21 WatchSource:0}: Error finding container 5af196b1dc0d416a7a6ae7e69e119e54ca2dba5a6d17a990a7c40a1d7ced6b21: Status 404 returned error can't find the container with id 5af196b1dc0d416a7a6ae7e69e119e54ca2dba5a6d17a990a7c40a1d7ced6b21 Oct 05 09:30:00 crc kubenswrapper[4935]: I1005 09:30:00.996160 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s"] Oct 05 09:30:01 crc kubenswrapper[4935]: I1005 09:30:01.477873 4935 generic.go:334] "Generic (PLEG): container finished" podID="8d654093-f8b9-4333-8949-52bd35aea8d1" containerID="cd27f460da8ed74c0ffd2dfc66aa734a65fa5751fd1a5f03a675f8f8fcbc59d6" exitCode=0 Oct 05 09:30:01 crc kubenswrapper[4935]: I1005 09:30:01.477937 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" event={"ID":"8d654093-f8b9-4333-8949-52bd35aea8d1","Type":"ContainerDied","Data":"cd27f460da8ed74c0ffd2dfc66aa734a65fa5751fd1a5f03a675f8f8fcbc59d6"} Oct 05 09:30:01 crc kubenswrapper[4935]: I1005 09:30:01.478284 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" event={"ID":"8d654093-f8b9-4333-8949-52bd35aea8d1","Type":"ContainerStarted","Data":"5af196b1dc0d416a7a6ae7e69e119e54ca2dba5a6d17a990a7c40a1d7ced6b21"} Oct 05 09:30:02 crc kubenswrapper[4935]: I1005 09:30:02.861859 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.026348 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d654093-f8b9-4333-8949-52bd35aea8d1-config-volume\") pod \"8d654093-f8b9-4333-8949-52bd35aea8d1\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.026481 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d654093-f8b9-4333-8949-52bd35aea8d1-secret-volume\") pod \"8d654093-f8b9-4333-8949-52bd35aea8d1\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.026958 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d654093-f8b9-4333-8949-52bd35aea8d1-config-volume" (OuterVolumeSpecName: "config-volume") pod "8d654093-f8b9-4333-8949-52bd35aea8d1" (UID: "8d654093-f8b9-4333-8949-52bd35aea8d1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.027197 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdsll\" (UniqueName: \"kubernetes.io/projected/8d654093-f8b9-4333-8949-52bd35aea8d1-kube-api-access-mdsll\") pod \"8d654093-f8b9-4333-8949-52bd35aea8d1\" (UID: \"8d654093-f8b9-4333-8949-52bd35aea8d1\") " Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.028261 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d654093-f8b9-4333-8949-52bd35aea8d1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.033571 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d654093-f8b9-4333-8949-52bd35aea8d1-kube-api-access-mdsll" (OuterVolumeSpecName: "kube-api-access-mdsll") pod "8d654093-f8b9-4333-8949-52bd35aea8d1" (UID: "8d654093-f8b9-4333-8949-52bd35aea8d1"). InnerVolumeSpecName "kube-api-access-mdsll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.041029 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d654093-f8b9-4333-8949-52bd35aea8d1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8d654093-f8b9-4333-8949-52bd35aea8d1" (UID: "8d654093-f8b9-4333-8949-52bd35aea8d1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.130667 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d654093-f8b9-4333-8949-52bd35aea8d1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.130708 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdsll\" (UniqueName: \"kubernetes.io/projected/8d654093-f8b9-4333-8949-52bd35aea8d1-kube-api-access-mdsll\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.500651 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" event={"ID":"8d654093-f8b9-4333-8949-52bd35aea8d1","Type":"ContainerDied","Data":"5af196b1dc0d416a7a6ae7e69e119e54ca2dba5a6d17a990a7c40a1d7ced6b21"} Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.500698 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5af196b1dc0d416a7a6ae7e69e119e54ca2dba5a6d17a990a7c40a1d7ced6b21" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.500749 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s" Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.972747 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq"] Oct 05 09:30:03 crc kubenswrapper[4935]: I1005 09:30:03.980813 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-sbhnq"] Oct 05 09:30:04 crc kubenswrapper[4935]: I1005 09:30:04.792832 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1" path="/var/lib/kubelet/pods/2ea58fa3-7301-4bbf-8bb1-38262bb5d5f1/volumes" Oct 05 09:30:22 crc kubenswrapper[4935]: I1005 09:30:22.956837 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-45hdj"] Oct 05 09:30:22 crc kubenswrapper[4935]: E1005 09:30:22.958007 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d654093-f8b9-4333-8949-52bd35aea8d1" containerName="collect-profiles" Oct 05 09:30:22 crc kubenswrapper[4935]: I1005 09:30:22.958024 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d654093-f8b9-4333-8949-52bd35aea8d1" containerName="collect-profiles" Oct 05 09:30:22 crc kubenswrapper[4935]: I1005 09:30:22.958283 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d654093-f8b9-4333-8949-52bd35aea8d1" containerName="collect-profiles" Oct 05 09:30:22 crc kubenswrapper[4935]: I1005 09:30:22.960517 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:22 crc kubenswrapper[4935]: I1005 09:30:22.985870 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-45hdj"] Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.104587 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5snkg\" (UniqueName: \"kubernetes.io/projected/e8685a2f-e577-42c0-959c-27d1a9229945-kube-api-access-5snkg\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.105099 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-utilities\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.105253 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-catalog-content\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.207566 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5snkg\" (UniqueName: \"kubernetes.io/projected/e8685a2f-e577-42c0-959c-27d1a9229945-kube-api-access-5snkg\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.207634 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-utilities\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.207695 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-catalog-content\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.208222 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-catalog-content\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.208466 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-utilities\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.762051 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5snkg\" (UniqueName: \"kubernetes.io/projected/e8685a2f-e577-42c0-959c-27d1a9229945-kube-api-access-5snkg\") pod \"redhat-marketplace-45hdj\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:23 crc kubenswrapper[4935]: I1005 09:30:23.888989 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:24 crc kubenswrapper[4935]: I1005 09:30:24.377242 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-45hdj"] Oct 05 09:30:24 crc kubenswrapper[4935]: I1005 09:30:24.742095 4935 generic.go:334] "Generic (PLEG): container finished" podID="e8685a2f-e577-42c0-959c-27d1a9229945" containerID="f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219" exitCode=0 Oct 05 09:30:24 crc kubenswrapper[4935]: I1005 09:30:24.742244 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45hdj" event={"ID":"e8685a2f-e577-42c0-959c-27d1a9229945","Type":"ContainerDied","Data":"f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219"} Oct 05 09:30:24 crc kubenswrapper[4935]: I1005 09:30:24.742402 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45hdj" event={"ID":"e8685a2f-e577-42c0-959c-27d1a9229945","Type":"ContainerStarted","Data":"696468ebf17a58ffcbcad08fefb6b6699bad2437e4f2d3e5e1e513c0fda369fb"} Oct 05 09:30:26 crc kubenswrapper[4935]: I1005 09:30:26.765751 4935 generic.go:334] "Generic (PLEG): container finished" podID="e8685a2f-e577-42c0-959c-27d1a9229945" containerID="07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31" exitCode=0 Oct 05 09:30:26 crc kubenswrapper[4935]: I1005 09:30:26.765858 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45hdj" event={"ID":"e8685a2f-e577-42c0-959c-27d1a9229945","Type":"ContainerDied","Data":"07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31"} Oct 05 09:30:27 crc kubenswrapper[4935]: I1005 09:30:27.778346 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45hdj" event={"ID":"e8685a2f-e577-42c0-959c-27d1a9229945","Type":"ContainerStarted","Data":"5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec"} Oct 05 09:30:27 crc kubenswrapper[4935]: I1005 09:30:27.811390 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-45hdj" podStartSLOduration=3.413829708 podStartE2EDuration="5.811362183s" podCreationTimestamp="2025-10-05 09:30:22 +0000 UTC" firstStartedPulling="2025-10-05 09:30:24.743902698 +0000 UTC m=+9458.626529148" lastFinishedPulling="2025-10-05 09:30:27.141435163 +0000 UTC m=+9461.024061623" observedRunningTime="2025-10-05 09:30:27.80326293 +0000 UTC m=+9461.685889410" watchObservedRunningTime="2025-10-05 09:30:27.811362183 +0000 UTC m=+9461.693988653" Oct 05 09:30:33 crc kubenswrapper[4935]: I1005 09:30:33.889554 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:33 crc kubenswrapper[4935]: I1005 09:30:33.890324 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:33 crc kubenswrapper[4935]: I1005 09:30:33.976142 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:34 crc kubenswrapper[4935]: I1005 09:30:34.950432 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:35 crc kubenswrapper[4935]: I1005 09:30:35.047923 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-45hdj"] Oct 05 09:30:36 crc kubenswrapper[4935]: I1005 09:30:36.888477 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-45hdj" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" containerName="registry-server" containerID="cri-o://5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec" gracePeriod=2 Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.492658 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.639809 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5snkg\" (UniqueName: \"kubernetes.io/projected/e8685a2f-e577-42c0-959c-27d1a9229945-kube-api-access-5snkg\") pod \"e8685a2f-e577-42c0-959c-27d1a9229945\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.640082 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-catalog-content\") pod \"e8685a2f-e577-42c0-959c-27d1a9229945\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.640136 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-utilities\") pod \"e8685a2f-e577-42c0-959c-27d1a9229945\" (UID: \"e8685a2f-e577-42c0-959c-27d1a9229945\") " Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.640975 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-utilities" (OuterVolumeSpecName: "utilities") pod "e8685a2f-e577-42c0-959c-27d1a9229945" (UID: "e8685a2f-e577-42c0-959c-27d1a9229945"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.647511 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8685a2f-e577-42c0-959c-27d1a9229945-kube-api-access-5snkg" (OuterVolumeSpecName: "kube-api-access-5snkg") pod "e8685a2f-e577-42c0-959c-27d1a9229945" (UID: "e8685a2f-e577-42c0-959c-27d1a9229945"). InnerVolumeSpecName "kube-api-access-5snkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.656683 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8685a2f-e577-42c0-959c-27d1a9229945" (UID: "e8685a2f-e577-42c0-959c-27d1a9229945"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.743531 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.743963 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8685a2f-e577-42c0-959c-27d1a9229945-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.744230 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5snkg\" (UniqueName: \"kubernetes.io/projected/e8685a2f-e577-42c0-959c-27d1a9229945-kube-api-access-5snkg\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.905294 4935 generic.go:334] "Generic (PLEG): container finished" podID="e8685a2f-e577-42c0-959c-27d1a9229945" containerID="5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec" exitCode=0 Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.905336 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45hdj" event={"ID":"e8685a2f-e577-42c0-959c-27d1a9229945","Type":"ContainerDied","Data":"5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec"} Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.905362 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45hdj" event={"ID":"e8685a2f-e577-42c0-959c-27d1a9229945","Type":"ContainerDied","Data":"696468ebf17a58ffcbcad08fefb6b6699bad2437e4f2d3e5e1e513c0fda369fb"} Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.905383 4935 scope.go:117] "RemoveContainer" containerID="5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.905428 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45hdj" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.945807 4935 scope.go:117] "RemoveContainer" containerID="07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31" Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.948344 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-45hdj"] Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.965671 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-45hdj"] Oct 05 09:30:37 crc kubenswrapper[4935]: I1005 09:30:37.976802 4935 scope.go:117] "RemoveContainer" containerID="f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219" Oct 05 09:30:38 crc kubenswrapper[4935]: I1005 09:30:38.053124 4935 scope.go:117] "RemoveContainer" containerID="5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec" Oct 05 09:30:38 crc kubenswrapper[4935]: E1005 09:30:38.054124 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec\": container with ID starting with 5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec not found: ID does not exist" containerID="5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec" Oct 05 09:30:38 crc kubenswrapper[4935]: I1005 09:30:38.054361 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec"} err="failed to get container status \"5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec\": rpc error: code = NotFound desc = could not find container \"5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec\": container with ID starting with 5b8a999ccd302a4df5d4c21847f71029e6f44f645d1a80f84f04e143f1c551ec not found: ID does not exist" Oct 05 09:30:38 crc kubenswrapper[4935]: I1005 09:30:38.054497 4935 scope.go:117] "RemoveContainer" containerID="07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31" Oct 05 09:30:38 crc kubenswrapper[4935]: E1005 09:30:38.055196 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31\": container with ID starting with 07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31 not found: ID does not exist" containerID="07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31" Oct 05 09:30:38 crc kubenswrapper[4935]: I1005 09:30:38.055238 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31"} err="failed to get container status \"07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31\": rpc error: code = NotFound desc = could not find container \"07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31\": container with ID starting with 07a004761fb60642fb5e5903ba343d5ac7a4227090dbf9fdbcf20be1ad458b31 not found: ID does not exist" Oct 05 09:30:38 crc kubenswrapper[4935]: I1005 09:30:38.055267 4935 scope.go:117] "RemoveContainer" containerID="f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219" Oct 05 09:30:38 crc kubenswrapper[4935]: E1005 09:30:38.055584 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219\": container with ID starting with f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219 not found: ID does not exist" containerID="f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219" Oct 05 09:30:38 crc kubenswrapper[4935]: I1005 09:30:38.055719 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219"} err="failed to get container status \"f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219\": rpc error: code = NotFound desc = could not find container \"f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219\": container with ID starting with f399d8e90177df1998b809e0497a17c91ed1b9dbbeda06b7f15dead113032219 not found: ID does not exist" Oct 05 09:30:38 crc kubenswrapper[4935]: I1005 09:30:38.790582 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" path="/var/lib/kubelet/pods/e8685a2f-e577-42c0-959c-27d1a9229945/volumes" Oct 05 09:30:44 crc kubenswrapper[4935]: I1005 09:30:44.289594 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:30:44 crc kubenswrapper[4935]: I1005 09:30:44.290210 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:31:03 crc kubenswrapper[4935]: I1005 09:31:03.666608 4935 scope.go:117] "RemoveContainer" containerID="3e46a7bdcd93059016a8b1ca22425a495c85a494178c9e958ed819f75ecb96c2" Oct 05 09:31:14 crc kubenswrapper[4935]: I1005 09:31:14.289454 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:31:14 crc kubenswrapper[4935]: I1005 09:31:14.290327 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.289209 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.289809 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.289853 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.290620 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"379bfbe0de420a386519e50fa2c6588e131783fe5dfe591ec5759067e2e26bbc"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.290683 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://379bfbe0de420a386519e50fa2c6588e131783fe5dfe591ec5759067e2e26bbc" gracePeriod=600 Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.649319 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="379bfbe0de420a386519e50fa2c6588e131783fe5dfe591ec5759067e2e26bbc" exitCode=0 Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.649379 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"379bfbe0de420a386519e50fa2c6588e131783fe5dfe591ec5759067e2e26bbc"} Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.649721 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9"} Oct 05 09:31:44 crc kubenswrapper[4935]: I1005 09:31:44.649743 4935 scope.go:117] "RemoveContainer" containerID="af4a645529dfa5f1cfcb671b0b679dea3f18c0ede895e97b586a08517c4f030b" Oct 05 09:33:03 crc kubenswrapper[4935]: I1005 09:33:03.798958 4935 scope.go:117] "RemoveContainer" containerID="3e89884e9368281e78a4197d702229690a14b051785d183918dd5ad55bc9ed51" Oct 05 09:33:03 crc kubenswrapper[4935]: I1005 09:33:03.842025 4935 scope.go:117] "RemoveContainer" containerID="a54253299ced6e2085abcf3fdd16e87a5937fdd339872ba042d8c2d91cb8ee5a" Oct 05 09:33:03 crc kubenswrapper[4935]: I1005 09:33:03.887791 4935 scope.go:117] "RemoveContainer" containerID="5bfb8ff2b2c76654cfe4f06053e2f97368c99078f0e73f0387753500632d639f" Oct 05 09:33:32 crc kubenswrapper[4935]: I1005 09:33:32.879099 4935 generic.go:334] "Generic (PLEG): container finished" podID="2bad5788-676a-436c-bfad-20d810bbd73e" containerID="3c9e98f7c5722d8cbfe652726daa0e19902562e2446d70d866f5069f50f976b1" exitCode=0 Oct 05 09:33:32 crc kubenswrapper[4935]: I1005 09:33:32.879195 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" event={"ID":"2bad5788-676a-436c-bfad-20d810bbd73e","Type":"ContainerDied","Data":"3c9e98f7c5722d8cbfe652726daa0e19902562e2446d70d866f5069f50f976b1"} Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.486814 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.581526 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-1\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.581617 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-0\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.581640 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-0\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.581702 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ssh-key\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.581801 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-1\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.581822 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swbrl\" (UniqueName: \"kubernetes.io/projected/2bad5788-676a-436c-bfad-20d810bbd73e-kube-api-access-swbrl\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.581864 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-1\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.581955 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ceph\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.582018 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-inventory\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.582056 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-0\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.582078 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-combined-ca-bundle\") pod \"2bad5788-676a-436c-bfad-20d810bbd73e\" (UID: \"2bad5788-676a-436c-bfad-20d810bbd73e\") " Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.596078 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.596696 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bad5788-676a-436c-bfad-20d810bbd73e-kube-api-access-swbrl" (OuterVolumeSpecName: "kube-api-access-swbrl") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "kube-api-access-swbrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.601863 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ceph" (OuterVolumeSpecName: "ceph") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.619851 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.632995 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-inventory" (OuterVolumeSpecName: "inventory") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.634500 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.638605 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.640429 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.640599 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.653724 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.656281 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "2bad5788-676a-436c-bfad-20d810bbd73e" (UID: "2bad5788-676a-436c-bfad-20d810bbd73e"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685002 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685046 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685059 4935 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685072 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685082 4935 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685093 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685102 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685109 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685117 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685127 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swbrl\" (UniqueName: \"kubernetes.io/projected/2bad5788-676a-436c-bfad-20d810bbd73e-kube-api-access-swbrl\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.685137 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2bad5788-676a-436c-bfad-20d810bbd73e-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.900985 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" event={"ID":"2bad5788-676a-436c-bfad-20d810bbd73e","Type":"ContainerDied","Data":"011487f79b8aef2161fc1c17f8eba732d33146f5f881e41c9d7d6d5730402920"} Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.901025 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-mrgrf" Oct 05 09:33:34 crc kubenswrapper[4935]: I1005 09:33:34.901034 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="011487f79b8aef2161fc1c17f8eba732d33146f5f881e41c9d7d6d5730402920" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.064742 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-c2gcm"] Oct 05 09:33:35 crc kubenswrapper[4935]: E1005 09:33:35.065514 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" containerName="extract-utilities" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.065537 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" containerName="extract-utilities" Oct 05 09:33:35 crc kubenswrapper[4935]: E1005 09:33:35.065576 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" containerName="registry-server" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.065585 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" containerName="registry-server" Oct 05 09:33:35 crc kubenswrapper[4935]: E1005 09:33:35.065601 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" containerName="extract-content" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.065611 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" containerName="extract-content" Oct 05 09:33:35 crc kubenswrapper[4935]: E1005 09:33:35.065633 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bad5788-676a-436c-bfad-20d810bbd73e" containerName="nova-cell1-openstack-openstack-cell1" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.065642 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bad5788-676a-436c-bfad-20d810bbd73e" containerName="nova-cell1-openstack-openstack-cell1" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.065920 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bad5788-676a-436c-bfad-20d810bbd73e" containerName="nova-cell1-openstack-openstack-cell1" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.065957 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8685a2f-e577-42c0-959c-27d1a9229945" containerName="registry-server" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.066795 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.070164 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.070182 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.070396 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.070549 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.071288 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.082993 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-c2gcm"] Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.096309 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vngd\" (UniqueName: \"kubernetes.io/projected/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-kube-api-access-7vngd\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.096392 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.096447 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceph\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.096581 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.096681 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.096726 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-inventory\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.096771 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ssh-key\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.096815 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.198674 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.198774 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceph\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.198799 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.199771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.199815 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-inventory\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.199831 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ssh-key\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.199865 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.199956 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vngd\" (UniqueName: \"kubernetes.io/projected/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-kube-api-access-7vngd\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.203226 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceph\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.203289 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-inventory\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.203654 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ssh-key\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.203844 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.204563 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.204712 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.205596 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.216651 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vngd\" (UniqueName: \"kubernetes.io/projected/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-kube-api-access-7vngd\") pod \"telemetry-openstack-openstack-cell1-c2gcm\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:35 crc kubenswrapper[4935]: I1005 09:33:35.441089 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:33:36 crc kubenswrapper[4935]: I1005 09:33:36.008366 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-c2gcm"] Oct 05 09:33:36 crc kubenswrapper[4935]: I1005 09:33:36.014711 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:33:36 crc kubenswrapper[4935]: I1005 09:33:36.926576 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" event={"ID":"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a","Type":"ContainerStarted","Data":"4eb00c43a8b31e0be2dc9967e4276e7d8b58dc68e204947c7a88a0bc34859062"} Oct 05 09:33:36 crc kubenswrapper[4935]: I1005 09:33:36.927395 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" event={"ID":"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a","Type":"ContainerStarted","Data":"17e515ab0decb2a7484afa5b3d8c64b81903741d3af1e60efaf712065ffe5a01"} Oct 05 09:33:36 crc kubenswrapper[4935]: I1005 09:33:36.948476 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" podStartSLOduration=1.5583741180000001 podStartE2EDuration="1.948455717s" podCreationTimestamp="2025-10-05 09:33:35 +0000 UTC" firstStartedPulling="2025-10-05 09:33:36.014466123 +0000 UTC m=+9649.897092593" lastFinishedPulling="2025-10-05 09:33:36.404547732 +0000 UTC m=+9650.287174192" observedRunningTime="2025-10-05 09:33:36.941551295 +0000 UTC m=+9650.824177755" watchObservedRunningTime="2025-10-05 09:33:36.948455717 +0000 UTC m=+9650.831082177" Oct 05 09:33:44 crc kubenswrapper[4935]: I1005 09:33:44.290804 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:33:44 crc kubenswrapper[4935]: I1005 09:33:44.291646 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:34:14 crc kubenswrapper[4935]: I1005 09:34:14.289162 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:34:14 crc kubenswrapper[4935]: I1005 09:34:14.289723 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.289551 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.290177 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.290237 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.290758 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.290834 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" gracePeriod=600 Oct 05 09:34:44 crc kubenswrapper[4935]: E1005 09:34:44.418427 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.739913 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" exitCode=0 Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.739956 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9"} Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.739990 4935 scope.go:117] "RemoveContainer" containerID="379bfbe0de420a386519e50fa2c6588e131783fe5dfe591ec5759067e2e26bbc" Oct 05 09:34:44 crc kubenswrapper[4935]: I1005 09:34:44.740570 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:34:44 crc kubenswrapper[4935]: E1005 09:34:44.740821 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:34:55 crc kubenswrapper[4935]: I1005 09:34:55.777755 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:34:55 crc kubenswrapper[4935]: E1005 09:34:55.778471 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:35:06 crc kubenswrapper[4935]: I1005 09:35:06.799386 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:35:06 crc kubenswrapper[4935]: E1005 09:35:06.800324 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:35:19 crc kubenswrapper[4935]: I1005 09:35:19.776861 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:35:19 crc kubenswrapper[4935]: E1005 09:35:19.777719 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:35:31 crc kubenswrapper[4935]: I1005 09:35:31.778979 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:35:31 crc kubenswrapper[4935]: E1005 09:35:31.780276 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:35:42 crc kubenswrapper[4935]: I1005 09:35:42.777851 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:35:42 crc kubenswrapper[4935]: E1005 09:35:42.779092 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:35:55 crc kubenswrapper[4935]: I1005 09:35:55.778257 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:35:55 crc kubenswrapper[4935]: E1005 09:35:55.779189 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:36:09 crc kubenswrapper[4935]: I1005 09:36:09.779529 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:36:09 crc kubenswrapper[4935]: E1005 09:36:09.780507 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.011685 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mmkgq"] Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.019859 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.025547 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mmkgq"] Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.113189 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt7tp\" (UniqueName: \"kubernetes.io/projected/301931b6-70bf-4092-807c-d76495381161-kube-api-access-vt7tp\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.113266 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-utilities\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.113339 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-catalog-content\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.215601 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-utilities\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.215692 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-catalog-content\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.215851 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt7tp\" (UniqueName: \"kubernetes.io/projected/301931b6-70bf-4092-807c-d76495381161-kube-api-access-vt7tp\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.216473 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-utilities\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.216663 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-catalog-content\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.262319 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt7tp\" (UniqueName: \"kubernetes.io/projected/301931b6-70bf-4092-807c-d76495381161-kube-api-access-vt7tp\") pod \"redhat-operators-mmkgq\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.377265 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.777977 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:36:21 crc kubenswrapper[4935]: E1005 09:36:21.778750 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.864379 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mmkgq"] Oct 05 09:36:21 crc kubenswrapper[4935]: I1005 09:36:21.910193 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmkgq" event={"ID":"301931b6-70bf-4092-807c-d76495381161","Type":"ContainerStarted","Data":"5367360fed014761b87c7f02764cf65d88410434c7f17e67e67d52cb9c2c2788"} Oct 05 09:36:22 crc kubenswrapper[4935]: I1005 09:36:22.923939 4935 generic.go:334] "Generic (PLEG): container finished" podID="301931b6-70bf-4092-807c-d76495381161" containerID="4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b" exitCode=0 Oct 05 09:36:22 crc kubenswrapper[4935]: I1005 09:36:22.924001 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmkgq" event={"ID":"301931b6-70bf-4092-807c-d76495381161","Type":"ContainerDied","Data":"4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b"} Oct 05 09:36:23 crc kubenswrapper[4935]: I1005 09:36:23.938913 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmkgq" event={"ID":"301931b6-70bf-4092-807c-d76495381161","Type":"ContainerStarted","Data":"5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391"} Oct 05 09:36:26 crc kubenswrapper[4935]: I1005 09:36:26.967669 4935 generic.go:334] "Generic (PLEG): container finished" podID="301931b6-70bf-4092-807c-d76495381161" containerID="5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391" exitCode=0 Oct 05 09:36:26 crc kubenswrapper[4935]: I1005 09:36:26.967766 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmkgq" event={"ID":"301931b6-70bf-4092-807c-d76495381161","Type":"ContainerDied","Data":"5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391"} Oct 05 09:36:27 crc kubenswrapper[4935]: I1005 09:36:27.981310 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmkgq" event={"ID":"301931b6-70bf-4092-807c-d76495381161","Type":"ContainerStarted","Data":"a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936"} Oct 05 09:36:27 crc kubenswrapper[4935]: I1005 09:36:27.999741 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mmkgq" podStartSLOduration=3.566107797 podStartE2EDuration="7.999724503s" podCreationTimestamp="2025-10-05 09:36:20 +0000 UTC" firstStartedPulling="2025-10-05 09:36:22.926325021 +0000 UTC m=+9816.808951521" lastFinishedPulling="2025-10-05 09:36:27.359941757 +0000 UTC m=+9821.242568227" observedRunningTime="2025-10-05 09:36:27.9980676 +0000 UTC m=+9821.880694060" watchObservedRunningTime="2025-10-05 09:36:27.999724503 +0000 UTC m=+9821.882350963" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.175831 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nl68v"] Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.178431 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.201695 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nl68v"] Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.269733 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-catalog-content\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.269817 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-utilities\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.270110 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8kgr\" (UniqueName: \"kubernetes.io/projected/8f6b2ab0-172d-476f-a644-6413df2466a1-kube-api-access-v8kgr\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.372922 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8kgr\" (UniqueName: \"kubernetes.io/projected/8f6b2ab0-172d-476f-a644-6413df2466a1-kube-api-access-v8kgr\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.373053 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-catalog-content\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.373080 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-utilities\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.373542 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-utilities\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.373691 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-catalog-content\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.407699 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8kgr\" (UniqueName: \"kubernetes.io/projected/8f6b2ab0-172d-476f-a644-6413df2466a1-kube-api-access-v8kgr\") pod \"community-operators-nl68v\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:28 crc kubenswrapper[4935]: I1005 09:36:28.511948 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:29 crc kubenswrapper[4935]: I1005 09:36:29.072915 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nl68v"] Oct 05 09:36:29 crc kubenswrapper[4935]: W1005 09:36:29.078943 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f6b2ab0_172d_476f_a644_6413df2466a1.slice/crio-432413b93dbaf8834e3e1e232f429e5c303ce3b120b0c29985813c6079e7b3fa WatchSource:0}: Error finding container 432413b93dbaf8834e3e1e232f429e5c303ce3b120b0c29985813c6079e7b3fa: Status 404 returned error can't find the container with id 432413b93dbaf8834e3e1e232f429e5c303ce3b120b0c29985813c6079e7b3fa Oct 05 09:36:29 crc kubenswrapper[4935]: I1005 09:36:29.998832 4935 generic.go:334] "Generic (PLEG): container finished" podID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerID="8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f" exitCode=0 Oct 05 09:36:29 crc kubenswrapper[4935]: I1005 09:36:29.998921 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl68v" event={"ID":"8f6b2ab0-172d-476f-a644-6413df2466a1","Type":"ContainerDied","Data":"8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f"} Oct 05 09:36:29 crc kubenswrapper[4935]: I1005 09:36:29.999158 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl68v" event={"ID":"8f6b2ab0-172d-476f-a644-6413df2466a1","Type":"ContainerStarted","Data":"432413b93dbaf8834e3e1e232f429e5c303ce3b120b0c29985813c6079e7b3fa"} Oct 05 09:36:31 crc kubenswrapper[4935]: I1005 09:36:31.026515 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl68v" event={"ID":"8f6b2ab0-172d-476f-a644-6413df2466a1","Type":"ContainerStarted","Data":"797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496"} Oct 05 09:36:31 crc kubenswrapper[4935]: I1005 09:36:31.378721 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:31 crc kubenswrapper[4935]: I1005 09:36:31.379965 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:32 crc kubenswrapper[4935]: I1005 09:36:32.433155 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mmkgq" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="registry-server" probeResult="failure" output=< Oct 05 09:36:32 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 09:36:32 crc kubenswrapper[4935]: > Oct 05 09:36:32 crc kubenswrapper[4935]: I1005 09:36:32.777650 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:36:32 crc kubenswrapper[4935]: E1005 09:36:32.777959 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:36:33 crc kubenswrapper[4935]: I1005 09:36:33.051094 4935 generic.go:334] "Generic (PLEG): container finished" podID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerID="797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496" exitCode=0 Oct 05 09:36:33 crc kubenswrapper[4935]: I1005 09:36:33.051143 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl68v" event={"ID":"8f6b2ab0-172d-476f-a644-6413df2466a1","Type":"ContainerDied","Data":"797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496"} Oct 05 09:36:34 crc kubenswrapper[4935]: I1005 09:36:34.065885 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl68v" event={"ID":"8f6b2ab0-172d-476f-a644-6413df2466a1","Type":"ContainerStarted","Data":"c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225"} Oct 05 09:36:34 crc kubenswrapper[4935]: I1005 09:36:34.093982 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nl68v" podStartSLOduration=2.404471599 podStartE2EDuration="6.093963005s" podCreationTimestamp="2025-10-05 09:36:28 +0000 UTC" firstStartedPulling="2025-10-05 09:36:30.000602037 +0000 UTC m=+9823.883228507" lastFinishedPulling="2025-10-05 09:36:33.690093443 +0000 UTC m=+9827.572719913" observedRunningTime="2025-10-05 09:36:34.083104529 +0000 UTC m=+9827.965731009" watchObservedRunningTime="2025-10-05 09:36:34.093963005 +0000 UTC m=+9827.976589465" Oct 05 09:36:38 crc kubenswrapper[4935]: I1005 09:36:38.512564 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:38 crc kubenswrapper[4935]: I1005 09:36:38.513381 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:38 crc kubenswrapper[4935]: I1005 09:36:38.569083 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:39 crc kubenswrapper[4935]: I1005 09:36:39.180227 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:39 crc kubenswrapper[4935]: I1005 09:36:39.247385 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nl68v"] Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.164653 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nl68v" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerName="registry-server" containerID="cri-o://c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225" gracePeriod=2 Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.462027 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.554128 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.741753 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.799343 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8kgr\" (UniqueName: \"kubernetes.io/projected/8f6b2ab0-172d-476f-a644-6413df2466a1-kube-api-access-v8kgr\") pod \"8f6b2ab0-172d-476f-a644-6413df2466a1\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.799422 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-catalog-content\") pod \"8f6b2ab0-172d-476f-a644-6413df2466a1\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.799604 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-utilities\") pod \"8f6b2ab0-172d-476f-a644-6413df2466a1\" (UID: \"8f6b2ab0-172d-476f-a644-6413df2466a1\") " Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.800692 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-utilities" (OuterVolumeSpecName: "utilities") pod "8f6b2ab0-172d-476f-a644-6413df2466a1" (UID: "8f6b2ab0-172d-476f-a644-6413df2466a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.806984 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f6b2ab0-172d-476f-a644-6413df2466a1-kube-api-access-v8kgr" (OuterVolumeSpecName: "kube-api-access-v8kgr") pod "8f6b2ab0-172d-476f-a644-6413df2466a1" (UID: "8f6b2ab0-172d-476f-a644-6413df2466a1"). InnerVolumeSpecName "kube-api-access-v8kgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.847591 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f6b2ab0-172d-476f-a644-6413df2466a1" (UID: "8f6b2ab0-172d-476f-a644-6413df2466a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.903087 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.903126 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6b2ab0-172d-476f-a644-6413df2466a1-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:36:41 crc kubenswrapper[4935]: I1005 09:36:41.903135 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8kgr\" (UniqueName: \"kubernetes.io/projected/8f6b2ab0-172d-476f-a644-6413df2466a1-kube-api-access-v8kgr\") on node \"crc\" DevicePath \"\"" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.177259 4935 generic.go:334] "Generic (PLEG): container finished" podID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerID="c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225" exitCode=0 Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.177321 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nl68v" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.177319 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl68v" event={"ID":"8f6b2ab0-172d-476f-a644-6413df2466a1","Type":"ContainerDied","Data":"c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225"} Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.177395 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl68v" event={"ID":"8f6b2ab0-172d-476f-a644-6413df2466a1","Type":"ContainerDied","Data":"432413b93dbaf8834e3e1e232f429e5c303ce3b120b0c29985813c6079e7b3fa"} Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.177430 4935 scope.go:117] "RemoveContainer" containerID="c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.216601 4935 scope.go:117] "RemoveContainer" containerID="797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.220505 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nl68v"] Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.230737 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nl68v"] Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.246443 4935 scope.go:117] "RemoveContainer" containerID="8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.290066 4935 scope.go:117] "RemoveContainer" containerID="c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225" Oct 05 09:36:42 crc kubenswrapper[4935]: E1005 09:36:42.290589 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225\": container with ID starting with c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225 not found: ID does not exist" containerID="c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.290720 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225"} err="failed to get container status \"c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225\": rpc error: code = NotFound desc = could not find container \"c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225\": container with ID starting with c9e755163595dcc038eac7ef88f1cc49f2dd8f34fda33a358e7ebbc54a77c225 not found: ID does not exist" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.290847 4935 scope.go:117] "RemoveContainer" containerID="797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496" Oct 05 09:36:42 crc kubenswrapper[4935]: E1005 09:36:42.291437 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496\": container with ID starting with 797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496 not found: ID does not exist" containerID="797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.291543 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496"} err="failed to get container status \"797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496\": rpc error: code = NotFound desc = could not find container \"797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496\": container with ID starting with 797923a0792fed3acf2c441ee548d2cf4a143433b2083bbf6d71c0e5ed50e496 not found: ID does not exist" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.291633 4935 scope.go:117] "RemoveContainer" containerID="8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f" Oct 05 09:36:42 crc kubenswrapper[4935]: E1005 09:36:42.292075 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f\": container with ID starting with 8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f not found: ID does not exist" containerID="8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.292380 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f"} err="failed to get container status \"8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f\": rpc error: code = NotFound desc = could not find container \"8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f\": container with ID starting with 8e0e722e1f1006ef6da71c42dab32e6ad65d344d81439f52a93c56a68a01030f not found: ID does not exist" Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.606102 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mmkgq"] Oct 05 09:36:42 crc kubenswrapper[4935]: I1005 09:36:42.799741 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" path="/var/lib/kubelet/pods/8f6b2ab0-172d-476f-a644-6413df2466a1/volumes" Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.192327 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mmkgq" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="registry-server" containerID="cri-o://a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936" gracePeriod=2 Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.737549 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.845469 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-catalog-content\") pod \"301931b6-70bf-4092-807c-d76495381161\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.845503 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-utilities\") pod \"301931b6-70bf-4092-807c-d76495381161\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.845681 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt7tp\" (UniqueName: \"kubernetes.io/projected/301931b6-70bf-4092-807c-d76495381161-kube-api-access-vt7tp\") pod \"301931b6-70bf-4092-807c-d76495381161\" (UID: \"301931b6-70bf-4092-807c-d76495381161\") " Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.846394 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-utilities" (OuterVolumeSpecName: "utilities") pod "301931b6-70bf-4092-807c-d76495381161" (UID: "301931b6-70bf-4092-807c-d76495381161"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.846911 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.852128 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/301931b6-70bf-4092-807c-d76495381161-kube-api-access-vt7tp" (OuterVolumeSpecName: "kube-api-access-vt7tp") pod "301931b6-70bf-4092-807c-d76495381161" (UID: "301931b6-70bf-4092-807c-d76495381161"). InnerVolumeSpecName "kube-api-access-vt7tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.927833 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "301931b6-70bf-4092-807c-d76495381161" (UID: "301931b6-70bf-4092-807c-d76495381161"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.948532 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt7tp\" (UniqueName: \"kubernetes.io/projected/301931b6-70bf-4092-807c-d76495381161-kube-api-access-vt7tp\") on node \"crc\" DevicePath \"\"" Oct 05 09:36:43 crc kubenswrapper[4935]: I1005 09:36:43.948560 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/301931b6-70bf-4092-807c-d76495381161-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.206195 4935 generic.go:334] "Generic (PLEG): container finished" podID="301931b6-70bf-4092-807c-d76495381161" containerID="a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936" exitCode=0 Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.206251 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmkgq" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.206256 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmkgq" event={"ID":"301931b6-70bf-4092-807c-d76495381161","Type":"ContainerDied","Data":"a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936"} Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.206401 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmkgq" event={"ID":"301931b6-70bf-4092-807c-d76495381161","Type":"ContainerDied","Data":"5367360fed014761b87c7f02764cf65d88410434c7f17e67e67d52cb9c2c2788"} Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.206426 4935 scope.go:117] "RemoveContainer" containerID="a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.248013 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mmkgq"] Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.248218 4935 scope.go:117] "RemoveContainer" containerID="5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.256597 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mmkgq"] Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.294681 4935 scope.go:117] "RemoveContainer" containerID="4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.339834 4935 scope.go:117] "RemoveContainer" containerID="a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936" Oct 05 09:36:44 crc kubenswrapper[4935]: E1005 09:36:44.340390 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936\": container with ID starting with a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936 not found: ID does not exist" containerID="a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.340446 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936"} err="failed to get container status \"a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936\": rpc error: code = NotFound desc = could not find container \"a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936\": container with ID starting with a999c8ec4762a22a311f094018f83404e5eb0650dcd14ed0da1721282c38e936 not found: ID does not exist" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.340479 4935 scope.go:117] "RemoveContainer" containerID="5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391" Oct 05 09:36:44 crc kubenswrapper[4935]: E1005 09:36:44.341004 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391\": container with ID starting with 5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391 not found: ID does not exist" containerID="5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.341045 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391"} err="failed to get container status \"5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391\": rpc error: code = NotFound desc = could not find container \"5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391\": container with ID starting with 5472b00376240650f943cd3d46cbf86f5016b0d7016b59db732176c895b39391 not found: ID does not exist" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.341093 4935 scope.go:117] "RemoveContainer" containerID="4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b" Oct 05 09:36:44 crc kubenswrapper[4935]: E1005 09:36:44.341739 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b\": container with ID starting with 4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b not found: ID does not exist" containerID="4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.341811 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b"} err="failed to get container status \"4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b\": rpc error: code = NotFound desc = could not find container \"4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b\": container with ID starting with 4446b11df74a565ec94e19da0d601ef47e38ef2bccbae82214162adb83a9104b not found: ID does not exist" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.778299 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:36:44 crc kubenswrapper[4935]: E1005 09:36:44.779064 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:36:44 crc kubenswrapper[4935]: I1005 09:36:44.804004 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="301931b6-70bf-4092-807c-d76495381161" path="/var/lib/kubelet/pods/301931b6-70bf-4092-807c-d76495381161/volumes" Oct 05 09:36:57 crc kubenswrapper[4935]: I1005 09:36:57.777467 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:36:57 crc kubenswrapper[4935]: E1005 09:36:57.778405 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:37:08 crc kubenswrapper[4935]: I1005 09:37:08.777465 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:37:08 crc kubenswrapper[4935]: E1005 09:37:08.778307 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:37:23 crc kubenswrapper[4935]: I1005 09:37:23.777610 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:37:23 crc kubenswrapper[4935]: E1005 09:37:23.778757 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:37:35 crc kubenswrapper[4935]: I1005 09:37:35.778112 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:37:35 crc kubenswrapper[4935]: E1005 09:37:35.779010 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:37:47 crc kubenswrapper[4935]: I1005 09:37:47.777688 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:37:47 crc kubenswrapper[4935]: E1005 09:37:47.778493 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:37:58 crc kubenswrapper[4935]: I1005 09:37:58.778284 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:37:58 crc kubenswrapper[4935]: E1005 09:37:58.779556 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:38:11 crc kubenswrapper[4935]: I1005 09:38:11.777206 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:38:11 crc kubenswrapper[4935]: E1005 09:38:11.778190 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:38:24 crc kubenswrapper[4935]: I1005 09:38:24.777653 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:38:24 crc kubenswrapper[4935]: E1005 09:38:24.778486 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:38:35 crc kubenswrapper[4935]: I1005 09:38:35.777996 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:38:35 crc kubenswrapper[4935]: E1005 09:38:35.797822 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:38:50 crc kubenswrapper[4935]: I1005 09:38:50.777056 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:38:50 crc kubenswrapper[4935]: E1005 09:38:50.778075 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.212697 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j4zhm"] Oct 05 09:39:01 crc kubenswrapper[4935]: E1005 09:39:01.213750 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="extract-content" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.213767 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="extract-content" Oct 05 09:39:01 crc kubenswrapper[4935]: E1005 09:39:01.213781 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerName="extract-utilities" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.213790 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerName="extract-utilities" Oct 05 09:39:01 crc kubenswrapper[4935]: E1005 09:39:01.213811 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerName="extract-content" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.213820 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerName="extract-content" Oct 05 09:39:01 crc kubenswrapper[4935]: E1005 09:39:01.213849 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="registry-server" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.213858 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="registry-server" Oct 05 09:39:01 crc kubenswrapper[4935]: E1005 09:39:01.213871 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerName="registry-server" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.213882 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerName="registry-server" Oct 05 09:39:01 crc kubenswrapper[4935]: E1005 09:39:01.213930 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="extract-utilities" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.213942 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="extract-utilities" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.214222 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="301931b6-70bf-4092-807c-d76495381161" containerName="registry-server" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.214261 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f6b2ab0-172d-476f-a644-6413df2466a1" containerName="registry-server" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.216334 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.251128 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j4zhm"] Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.306224 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-utilities\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.306306 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghx9\" (UniqueName: \"kubernetes.io/projected/3c83f1d1-f8bb-4751-b4c5-438a12f343da-kube-api-access-rghx9\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.306409 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-catalog-content\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.408876 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-utilities\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.408961 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghx9\" (UniqueName: \"kubernetes.io/projected/3c83f1d1-f8bb-4751-b4c5-438a12f343da-kube-api-access-rghx9\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.409008 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-catalog-content\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.409677 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-utilities\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.409874 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-catalog-content\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.665224 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghx9\" (UniqueName: \"kubernetes.io/projected/3c83f1d1-f8bb-4751-b4c5-438a12f343da-kube-api-access-rghx9\") pod \"certified-operators-j4zhm\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:01 crc kubenswrapper[4935]: I1005 09:39:01.889815 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:02 crc kubenswrapper[4935]: I1005 09:39:02.316007 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j4zhm"] Oct 05 09:39:02 crc kubenswrapper[4935]: I1005 09:39:02.868757 4935 generic.go:334] "Generic (PLEG): container finished" podID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerID="116e089247af7fd7e057febc87e60a5fe5ec3c3d8325032931041335d3c50e20" exitCode=0 Oct 05 09:39:02 crc kubenswrapper[4935]: I1005 09:39:02.868819 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j4zhm" event={"ID":"3c83f1d1-f8bb-4751-b4c5-438a12f343da","Type":"ContainerDied","Data":"116e089247af7fd7e057febc87e60a5fe5ec3c3d8325032931041335d3c50e20"} Oct 05 09:39:02 crc kubenswrapper[4935]: I1005 09:39:02.868851 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j4zhm" event={"ID":"3c83f1d1-f8bb-4751-b4c5-438a12f343da","Type":"ContainerStarted","Data":"53b6f512d5a312ac1b7ecca13583d9e1ffc94d8d8f0551fc845721b1a7dff4f1"} Oct 05 09:39:02 crc kubenswrapper[4935]: I1005 09:39:02.871784 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:39:03 crc kubenswrapper[4935]: I1005 09:39:03.777815 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:39:03 crc kubenswrapper[4935]: E1005 09:39:03.778528 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:39:04 crc kubenswrapper[4935]: I1005 09:39:04.895191 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j4zhm" event={"ID":"3c83f1d1-f8bb-4751-b4c5-438a12f343da","Type":"ContainerStarted","Data":"df7181d713ce38a7891aa918028fc6c74ecc86bd6e14e6d3de275de8fe14ad37"} Oct 05 09:39:05 crc kubenswrapper[4935]: I1005 09:39:05.911685 4935 generic.go:334] "Generic (PLEG): container finished" podID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerID="df7181d713ce38a7891aa918028fc6c74ecc86bd6e14e6d3de275de8fe14ad37" exitCode=0 Oct 05 09:39:05 crc kubenswrapper[4935]: I1005 09:39:05.911783 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j4zhm" event={"ID":"3c83f1d1-f8bb-4751-b4c5-438a12f343da","Type":"ContainerDied","Data":"df7181d713ce38a7891aa918028fc6c74ecc86bd6e14e6d3de275de8fe14ad37"} Oct 05 09:39:06 crc kubenswrapper[4935]: I1005 09:39:06.930522 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j4zhm" event={"ID":"3c83f1d1-f8bb-4751-b4c5-438a12f343da","Type":"ContainerStarted","Data":"9b5dac795dd87e49b5206519fe4d7b51678120b59711c744bffec3557f31a1a3"} Oct 05 09:39:06 crc kubenswrapper[4935]: I1005 09:39:06.957007 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j4zhm" podStartSLOduration=2.518673243 podStartE2EDuration="5.956993241s" podCreationTimestamp="2025-10-05 09:39:01 +0000 UTC" firstStartedPulling="2025-10-05 09:39:02.871508782 +0000 UTC m=+9976.754135242" lastFinishedPulling="2025-10-05 09:39:06.30982878 +0000 UTC m=+9980.192455240" observedRunningTime="2025-10-05 09:39:06.954064174 +0000 UTC m=+9980.836690634" watchObservedRunningTime="2025-10-05 09:39:06.956993241 +0000 UTC m=+9980.839619701" Oct 05 09:39:11 crc kubenswrapper[4935]: I1005 09:39:11.890955 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:11 crc kubenswrapper[4935]: I1005 09:39:11.891481 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:11 crc kubenswrapper[4935]: I1005 09:39:11.955075 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:12 crc kubenswrapper[4935]: I1005 09:39:12.029651 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:12 crc kubenswrapper[4935]: I1005 09:39:12.201233 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j4zhm"] Oct 05 09:39:14 crc kubenswrapper[4935]: I1005 09:39:14.007080 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j4zhm" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerName="registry-server" containerID="cri-o://9b5dac795dd87e49b5206519fe4d7b51678120b59711c744bffec3557f31a1a3" gracePeriod=2 Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.028532 4935 generic.go:334] "Generic (PLEG): container finished" podID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerID="9b5dac795dd87e49b5206519fe4d7b51678120b59711c744bffec3557f31a1a3" exitCode=0 Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.028723 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j4zhm" event={"ID":"3c83f1d1-f8bb-4751-b4c5-438a12f343da","Type":"ContainerDied","Data":"9b5dac795dd87e49b5206519fe4d7b51678120b59711c744bffec3557f31a1a3"} Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.382290 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.532052 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rghx9\" (UniqueName: \"kubernetes.io/projected/3c83f1d1-f8bb-4751-b4c5-438a12f343da-kube-api-access-rghx9\") pod \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.532209 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-catalog-content\") pod \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.532321 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-utilities\") pod \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\" (UID: \"3c83f1d1-f8bb-4751-b4c5-438a12f343da\") " Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.533413 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-utilities" (OuterVolumeSpecName: "utilities") pod "3c83f1d1-f8bb-4751-b4c5-438a12f343da" (UID: "3c83f1d1-f8bb-4751-b4c5-438a12f343da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.540371 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c83f1d1-f8bb-4751-b4c5-438a12f343da-kube-api-access-rghx9" (OuterVolumeSpecName: "kube-api-access-rghx9") pod "3c83f1d1-f8bb-4751-b4c5-438a12f343da" (UID: "3c83f1d1-f8bb-4751-b4c5-438a12f343da"). InnerVolumeSpecName "kube-api-access-rghx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.579130 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c83f1d1-f8bb-4751-b4c5-438a12f343da" (UID: "3c83f1d1-f8bb-4751-b4c5-438a12f343da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.635386 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.635424 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rghx9\" (UniqueName: \"kubernetes.io/projected/3c83f1d1-f8bb-4751-b4c5-438a12f343da-kube-api-access-rghx9\") on node \"crc\" DevicePath \"\"" Oct 05 09:39:15 crc kubenswrapper[4935]: I1005 09:39:15.635441 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c83f1d1-f8bb-4751-b4c5-438a12f343da-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:39:16 crc kubenswrapper[4935]: I1005 09:39:16.046609 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j4zhm" event={"ID":"3c83f1d1-f8bb-4751-b4c5-438a12f343da","Type":"ContainerDied","Data":"53b6f512d5a312ac1b7ecca13583d9e1ffc94d8d8f0551fc845721b1a7dff4f1"} Oct 05 09:39:16 crc kubenswrapper[4935]: I1005 09:39:16.046664 4935 scope.go:117] "RemoveContainer" containerID="9b5dac795dd87e49b5206519fe4d7b51678120b59711c744bffec3557f31a1a3" Oct 05 09:39:16 crc kubenswrapper[4935]: I1005 09:39:16.048149 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j4zhm" Oct 05 09:39:16 crc kubenswrapper[4935]: I1005 09:39:16.093422 4935 scope.go:117] "RemoveContainer" containerID="df7181d713ce38a7891aa918028fc6c74ecc86bd6e14e6d3de275de8fe14ad37" Oct 05 09:39:16 crc kubenswrapper[4935]: I1005 09:39:16.105404 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j4zhm"] Oct 05 09:39:16 crc kubenswrapper[4935]: I1005 09:39:16.129324 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j4zhm"] Oct 05 09:39:16 crc kubenswrapper[4935]: I1005 09:39:16.132980 4935 scope.go:117] "RemoveContainer" containerID="116e089247af7fd7e057febc87e60a5fe5ec3c3d8325032931041335d3c50e20" Oct 05 09:39:16 crc kubenswrapper[4935]: I1005 09:39:16.793276 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" path="/var/lib/kubelet/pods/3c83f1d1-f8bb-4751-b4c5-438a12f343da/volumes" Oct 05 09:39:17 crc kubenswrapper[4935]: I1005 09:39:17.778086 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:39:17 crc kubenswrapper[4935]: E1005 09:39:17.779510 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:39:31 crc kubenswrapper[4935]: I1005 09:39:31.776940 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:39:31 crc kubenswrapper[4935]: E1005 09:39:31.777542 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:39:42 crc kubenswrapper[4935]: I1005 09:39:42.778047 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:39:42 crc kubenswrapper[4935]: E1005 09:39:42.780922 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:39:54 crc kubenswrapper[4935]: I1005 09:39:54.777616 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:39:55 crc kubenswrapper[4935]: I1005 09:39:55.556080 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"c31b5e83aad13ab5aad5147c3c65b8e692b7949147dda7547b1fcc298444ac8b"} Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.607986 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cfcfr"] Oct 05 09:40:37 crc kubenswrapper[4935]: E1005 09:40:37.609772 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerName="registry-server" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.609809 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerName="registry-server" Oct 05 09:40:37 crc kubenswrapper[4935]: E1005 09:40:37.609837 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerName="extract-content" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.609853 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerName="extract-content" Oct 05 09:40:37 crc kubenswrapper[4935]: E1005 09:40:37.609957 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerName="extract-utilities" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.609978 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerName="extract-utilities" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.610531 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c83f1d1-f8bb-4751-b4c5-438a12f343da" containerName="registry-server" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.614640 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.623004 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfcfr"] Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.761669 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-catalog-content\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.761722 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbdz4\" (UniqueName: \"kubernetes.io/projected/23227e3f-759e-4708-bc40-cf295d5203af-kube-api-access-gbdz4\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.762179 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-utilities\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.863615 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbdz4\" (UniqueName: \"kubernetes.io/projected/23227e3f-759e-4708-bc40-cf295d5203af-kube-api-access-gbdz4\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.864154 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-utilities\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.864467 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-catalog-content\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.865267 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-utilities\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.865345 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-catalog-content\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.903129 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbdz4\" (UniqueName: \"kubernetes.io/projected/23227e3f-759e-4708-bc40-cf295d5203af-kube-api-access-gbdz4\") pod \"redhat-marketplace-cfcfr\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:37 crc kubenswrapper[4935]: I1005 09:40:37.955004 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:38 crc kubenswrapper[4935]: I1005 09:40:38.467692 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfcfr"] Oct 05 09:40:39 crc kubenswrapper[4935]: I1005 09:40:39.110544 4935 generic.go:334] "Generic (PLEG): container finished" podID="23227e3f-759e-4708-bc40-cf295d5203af" containerID="d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032" exitCode=0 Oct 05 09:40:39 crc kubenswrapper[4935]: I1005 09:40:39.110626 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfcfr" event={"ID":"23227e3f-759e-4708-bc40-cf295d5203af","Type":"ContainerDied","Data":"d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032"} Oct 05 09:40:39 crc kubenswrapper[4935]: I1005 09:40:39.111033 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfcfr" event={"ID":"23227e3f-759e-4708-bc40-cf295d5203af","Type":"ContainerStarted","Data":"6efa368a3a15cdd89e1295ff04cedd134706a607b2f2ad41cbf47f55a224aeb8"} Oct 05 09:40:41 crc kubenswrapper[4935]: I1005 09:40:41.138747 4935 generic.go:334] "Generic (PLEG): container finished" podID="23227e3f-759e-4708-bc40-cf295d5203af" containerID="b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b" exitCode=0 Oct 05 09:40:41 crc kubenswrapper[4935]: I1005 09:40:41.138852 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfcfr" event={"ID":"23227e3f-759e-4708-bc40-cf295d5203af","Type":"ContainerDied","Data":"b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b"} Oct 05 09:40:42 crc kubenswrapper[4935]: I1005 09:40:42.154284 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfcfr" event={"ID":"23227e3f-759e-4708-bc40-cf295d5203af","Type":"ContainerStarted","Data":"4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09"} Oct 05 09:40:42 crc kubenswrapper[4935]: I1005 09:40:42.175097 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cfcfr" podStartSLOduration=2.6699525570000002 podStartE2EDuration="5.175079133s" podCreationTimestamp="2025-10-05 09:40:37 +0000 UTC" firstStartedPulling="2025-10-05 09:40:39.113247936 +0000 UTC m=+10072.995874406" lastFinishedPulling="2025-10-05 09:40:41.618374522 +0000 UTC m=+10075.501000982" observedRunningTime="2025-10-05 09:40:42.171236182 +0000 UTC m=+10076.053862642" watchObservedRunningTime="2025-10-05 09:40:42.175079133 +0000 UTC m=+10076.057705593" Oct 05 09:40:44 crc kubenswrapper[4935]: I1005 09:40:44.181354 4935 generic.go:334] "Generic (PLEG): container finished" podID="5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" containerID="4eb00c43a8b31e0be2dc9967e4276e7d8b58dc68e204947c7a88a0bc34859062" exitCode=0 Oct 05 09:40:44 crc kubenswrapper[4935]: I1005 09:40:44.181463 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" event={"ID":"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a","Type":"ContainerDied","Data":"4eb00c43a8b31e0be2dc9967e4276e7d8b58dc68e204947c7a88a0bc34859062"} Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.713721 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.846616 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vngd\" (UniqueName: \"kubernetes.io/projected/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-kube-api-access-7vngd\") pod \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.846710 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-inventory\") pod \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.846836 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-0\") pod \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.846948 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceph\") pod \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.846979 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-1\") pod \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.847094 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-telemetry-combined-ca-bundle\") pod \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.847154 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ssh-key\") pod \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.847403 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-2\") pod \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\" (UID: \"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a\") " Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.852845 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-kube-api-access-7vngd" (OuterVolumeSpecName: "kube-api-access-7vngd") pod "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" (UID: "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a"). InnerVolumeSpecName "kube-api-access-7vngd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.860696 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceph" (OuterVolumeSpecName: "ceph") pod "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" (UID: "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.861820 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" (UID: "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.876390 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-inventory" (OuterVolumeSpecName: "inventory") pod "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" (UID: "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.877987 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" (UID: "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.880677 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" (UID: "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.915115 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" (UID: "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.916099 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" (UID: "5603bea1-f3c6-42b1-9f3c-5f4f10b4324a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.950527 4935 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.950571 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vngd\" (UniqueName: \"kubernetes.io/projected/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-kube-api-access-7vngd\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.950584 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.950597 4935 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.950608 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.950617 4935 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.950628 4935 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:45 crc kubenswrapper[4935]: I1005 09:40:45.950638 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5603bea1-f3c6-42b1-9f3c-5f4f10b4324a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.210806 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" event={"ID":"5603bea1-f3c6-42b1-9f3c-5f4f10b4324a","Type":"ContainerDied","Data":"17e515ab0decb2a7484afa5b3d8c64b81903741d3af1e60efaf712065ffe5a01"} Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.210857 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17e515ab0decb2a7484afa5b3d8c64b81903741d3af1e60efaf712065ffe5a01" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.211002 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-c2gcm" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.342027 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qqmlf"] Oct 05 09:40:46 crc kubenswrapper[4935]: E1005 09:40:46.342778 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" containerName="telemetry-openstack-openstack-cell1" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.342805 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" containerName="telemetry-openstack-openstack-cell1" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.343181 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5603bea1-f3c6-42b1-9f3c-5f4f10b4324a" containerName="telemetry-openstack-openstack-cell1" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.344125 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.348004 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.348026 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.348317 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.349139 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.351625 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.355037 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qqmlf"] Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.466337 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.466649 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz9np\" (UniqueName: \"kubernetes.io/projected/fcf454be-3f70-46c3-904b-dd763a1b35a7-kube-api-access-jz9np\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.466858 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.467003 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.467176 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.467324 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.569392 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.570562 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.570826 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.571049 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz9np\" (UniqueName: \"kubernetes.io/projected/fcf454be-3f70-46c3-904b-dd763a1b35a7-kube-api-access-jz9np\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.571284 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.571453 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.974029 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.976507 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.976548 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.976849 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.979115 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz9np\" (UniqueName: \"kubernetes.io/projected/fcf454be-3f70-46c3-904b-dd763a1b35a7-kube-api-access-jz9np\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:46 crc kubenswrapper[4935]: I1005 09:40:46.986931 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qqmlf\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:47 crc kubenswrapper[4935]: I1005 09:40:47.275745 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:40:47 crc kubenswrapper[4935]: I1005 09:40:47.285065 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:40:47 crc kubenswrapper[4935]: I1005 09:40:47.833210 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qqmlf"] Oct 05 09:40:47 crc kubenswrapper[4935]: I1005 09:40:47.955848 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:47 crc kubenswrapper[4935]: I1005 09:40:47.956154 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:48 crc kubenswrapper[4935]: I1005 09:40:48.007217 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:48 crc kubenswrapper[4935]: I1005 09:40:48.228909 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" event={"ID":"fcf454be-3f70-46c3-904b-dd763a1b35a7","Type":"ContainerStarted","Data":"d72ec39f49185ba661bc13546776d36cde5a562c0cf6dac95446f7b1d94bbc89"} Oct 05 09:40:48 crc kubenswrapper[4935]: I1005 09:40:48.274309 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:48 crc kubenswrapper[4935]: I1005 09:40:48.303874 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:40:48 crc kubenswrapper[4935]: I1005 09:40:48.325876 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfcfr"] Oct 05 09:40:49 crc kubenswrapper[4935]: I1005 09:40:49.241095 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" event={"ID":"fcf454be-3f70-46c3-904b-dd763a1b35a7","Type":"ContainerStarted","Data":"0f809edeb3f3be8aabe0382dc5c2380871b87a086284d235c6dca59d9169dbbc"} Oct 05 09:40:49 crc kubenswrapper[4935]: I1005 09:40:49.269407 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" podStartSLOduration=2.80454714 podStartE2EDuration="3.269386816s" podCreationTimestamp="2025-10-05 09:40:46 +0000 UTC" firstStartedPulling="2025-10-05 09:40:47.836863741 +0000 UTC m=+10081.719490201" lastFinishedPulling="2025-10-05 09:40:48.301703417 +0000 UTC m=+10082.184329877" observedRunningTime="2025-10-05 09:40:49.259763273 +0000 UTC m=+10083.142389743" watchObservedRunningTime="2025-10-05 09:40:49.269386816 +0000 UTC m=+10083.152013276" Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.250964 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cfcfr" podUID="23227e3f-759e-4708-bc40-cf295d5203af" containerName="registry-server" containerID="cri-o://4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09" gracePeriod=2 Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.746405 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.864104 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbdz4\" (UniqueName: \"kubernetes.io/projected/23227e3f-759e-4708-bc40-cf295d5203af-kube-api-access-gbdz4\") pod \"23227e3f-759e-4708-bc40-cf295d5203af\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.864435 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-catalog-content\") pod \"23227e3f-759e-4708-bc40-cf295d5203af\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.864571 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-utilities\") pod \"23227e3f-759e-4708-bc40-cf295d5203af\" (UID: \"23227e3f-759e-4708-bc40-cf295d5203af\") " Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.867227 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-utilities" (OuterVolumeSpecName: "utilities") pod "23227e3f-759e-4708-bc40-cf295d5203af" (UID: "23227e3f-759e-4708-bc40-cf295d5203af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.869626 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23227e3f-759e-4708-bc40-cf295d5203af-kube-api-access-gbdz4" (OuterVolumeSpecName: "kube-api-access-gbdz4") pod "23227e3f-759e-4708-bc40-cf295d5203af" (UID: "23227e3f-759e-4708-bc40-cf295d5203af"). InnerVolumeSpecName "kube-api-access-gbdz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.892558 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23227e3f-759e-4708-bc40-cf295d5203af" (UID: "23227e3f-759e-4708-bc40-cf295d5203af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.967602 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbdz4\" (UniqueName: \"kubernetes.io/projected/23227e3f-759e-4708-bc40-cf295d5203af-kube-api-access-gbdz4\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.967827 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:50 crc kubenswrapper[4935]: I1005 09:40:50.967927 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23227e3f-759e-4708-bc40-cf295d5203af-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.267996 4935 generic.go:334] "Generic (PLEG): container finished" podID="23227e3f-759e-4708-bc40-cf295d5203af" containerID="4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09" exitCode=0 Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.268052 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfcfr" event={"ID":"23227e3f-759e-4708-bc40-cf295d5203af","Type":"ContainerDied","Data":"4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09"} Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.268146 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfcfr" event={"ID":"23227e3f-759e-4708-bc40-cf295d5203af","Type":"ContainerDied","Data":"6efa368a3a15cdd89e1295ff04cedd134706a607b2f2ad41cbf47f55a224aeb8"} Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.268178 4935 scope.go:117] "RemoveContainer" containerID="4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.268209 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cfcfr" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.315826 4935 scope.go:117] "RemoveContainer" containerID="b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.330943 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfcfr"] Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.343181 4935 scope.go:117] "RemoveContainer" containerID="d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.344208 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfcfr"] Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.380763 4935 scope.go:117] "RemoveContainer" containerID="4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09" Oct 05 09:40:51 crc kubenswrapper[4935]: E1005 09:40:51.381424 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09\": container with ID starting with 4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09 not found: ID does not exist" containerID="4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.381476 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09"} err="failed to get container status \"4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09\": rpc error: code = NotFound desc = could not find container \"4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09\": container with ID starting with 4bf261e64888ac5e2fd06e3ac73fb8d5b6312ac5e0c083af38dbf702aabece09 not found: ID does not exist" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.381505 4935 scope.go:117] "RemoveContainer" containerID="b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b" Oct 05 09:40:51 crc kubenswrapper[4935]: E1005 09:40:51.381967 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b\": container with ID starting with b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b not found: ID does not exist" containerID="b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.382000 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b"} err="failed to get container status \"b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b\": rpc error: code = NotFound desc = could not find container \"b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b\": container with ID starting with b942260e3230e9e2795b9ad7c66bb1896137355b369194889a550de2d024e90b not found: ID does not exist" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.382021 4935 scope.go:117] "RemoveContainer" containerID="d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032" Oct 05 09:40:51 crc kubenswrapper[4935]: E1005 09:40:51.382729 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032\": container with ID starting with d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032 not found: ID does not exist" containerID="d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032" Oct 05 09:40:51 crc kubenswrapper[4935]: I1005 09:40:51.382777 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032"} err="failed to get container status \"d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032\": rpc error: code = NotFound desc = could not find container \"d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032\": container with ID starting with d666853576d7381a903f917032b2a633db70b425c6b2e43df871a6fb2041b032 not found: ID does not exist" Oct 05 09:40:52 crc kubenswrapper[4935]: I1005 09:40:52.791461 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23227e3f-759e-4708-bc40-cf295d5203af" path="/var/lib/kubelet/pods/23227e3f-759e-4708-bc40-cf295d5203af/volumes" Oct 05 09:41:35 crc kubenswrapper[4935]: I1005 09:41:35.770763 4935 generic.go:334] "Generic (PLEG): container finished" podID="fcf454be-3f70-46c3-904b-dd763a1b35a7" containerID="0f809edeb3f3be8aabe0382dc5c2380871b87a086284d235c6dca59d9169dbbc" exitCode=0 Oct 05 09:41:35 crc kubenswrapper[4935]: I1005 09:41:35.770857 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" event={"ID":"fcf454be-3f70-46c3-904b-dd763a1b35a7","Type":"ContainerDied","Data":"0f809edeb3f3be8aabe0382dc5c2380871b87a086284d235c6dca59d9169dbbc"} Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.290112 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.401580 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ceph\") pod \"fcf454be-3f70-46c3-904b-dd763a1b35a7\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.401781 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-inventory\") pod \"fcf454be-3f70-46c3-904b-dd763a1b35a7\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.401842 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz9np\" (UniqueName: \"kubernetes.io/projected/fcf454be-3f70-46c3-904b-dd763a1b35a7-kube-api-access-jz9np\") pod \"fcf454be-3f70-46c3-904b-dd763a1b35a7\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.401902 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-combined-ca-bundle\") pod \"fcf454be-3f70-46c3-904b-dd763a1b35a7\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.401928 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-agent-neutron-config-0\") pod \"fcf454be-3f70-46c3-904b-dd763a1b35a7\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.401947 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ssh-key\") pod \"fcf454be-3f70-46c3-904b-dd763a1b35a7\" (UID: \"fcf454be-3f70-46c3-904b-dd763a1b35a7\") " Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.407284 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcf454be-3f70-46c3-904b-dd763a1b35a7-kube-api-access-jz9np" (OuterVolumeSpecName: "kube-api-access-jz9np") pod "fcf454be-3f70-46c3-904b-dd763a1b35a7" (UID: "fcf454be-3f70-46c3-904b-dd763a1b35a7"). InnerVolumeSpecName "kube-api-access-jz9np". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.407530 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ceph" (OuterVolumeSpecName: "ceph") pod "fcf454be-3f70-46c3-904b-dd763a1b35a7" (UID: "fcf454be-3f70-46c3-904b-dd763a1b35a7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.409995 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "fcf454be-3f70-46c3-904b-dd763a1b35a7" (UID: "fcf454be-3f70-46c3-904b-dd763a1b35a7"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.437286 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "fcf454be-3f70-46c3-904b-dd763a1b35a7" (UID: "fcf454be-3f70-46c3-904b-dd763a1b35a7"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.437413 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-inventory" (OuterVolumeSpecName: "inventory") pod "fcf454be-3f70-46c3-904b-dd763a1b35a7" (UID: "fcf454be-3f70-46c3-904b-dd763a1b35a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.437711 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fcf454be-3f70-46c3-904b-dd763a1b35a7" (UID: "fcf454be-3f70-46c3-904b-dd763a1b35a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.504872 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.504920 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.504934 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz9np\" (UniqueName: \"kubernetes.io/projected/fcf454be-3f70-46c3-904b-dd763a1b35a7-kube-api-access-jz9np\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.504944 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.504953 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.504964 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf454be-3f70-46c3-904b-dd763a1b35a7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.793935 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" event={"ID":"fcf454be-3f70-46c3-904b-dd763a1b35a7","Type":"ContainerDied","Data":"d72ec39f49185ba661bc13546776d36cde5a562c0cf6dac95446f7b1d94bbc89"} Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.793988 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d72ec39f49185ba661bc13546776d36cde5a562c0cf6dac95446f7b1d94bbc89" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.794050 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qqmlf" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.911112 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w"] Oct 05 09:41:37 crc kubenswrapper[4935]: E1005 09:41:37.912349 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23227e3f-759e-4708-bc40-cf295d5203af" containerName="extract-content" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.912375 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="23227e3f-759e-4708-bc40-cf295d5203af" containerName="extract-content" Oct 05 09:41:37 crc kubenswrapper[4935]: E1005 09:41:37.912459 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23227e3f-759e-4708-bc40-cf295d5203af" containerName="registry-server" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.912472 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="23227e3f-759e-4708-bc40-cf295d5203af" containerName="registry-server" Oct 05 09:41:37 crc kubenswrapper[4935]: E1005 09:41:37.912559 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23227e3f-759e-4708-bc40-cf295d5203af" containerName="extract-utilities" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.912574 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="23227e3f-759e-4708-bc40-cf295d5203af" containerName="extract-utilities" Oct 05 09:41:37 crc kubenswrapper[4935]: E1005 09:41:37.912614 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf454be-3f70-46c3-904b-dd763a1b35a7" containerName="neutron-sriov-openstack-openstack-cell1" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.912624 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf454be-3f70-46c3-904b-dd763a1b35a7" containerName="neutron-sriov-openstack-openstack-cell1" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.913269 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="23227e3f-759e-4708-bc40-cf295d5203af" containerName="registry-server" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.913330 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf454be-3f70-46c3-904b-dd763a1b35a7" containerName="neutron-sriov-openstack-openstack-cell1" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.922876 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.927669 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.927776 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.928160 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.928354 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.928584 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:41:37 crc kubenswrapper[4935]: I1005 09:41:37.935110 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w"] Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.022663 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.022740 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.022779 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.022852 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.022971 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktsb6\" (UniqueName: \"kubernetes.io/projected/5751e378-39fa-4bf1-98b1-b1242d642c2c-kube-api-access-ktsb6\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.023024 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.124910 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.125110 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktsb6\" (UniqueName: \"kubernetes.io/projected/5751e378-39fa-4bf1-98b1-b1242d642c2c-kube-api-access-ktsb6\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.125157 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.125881 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.125955 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.125991 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.129623 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.129723 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.129785 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.130548 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.132189 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.140529 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktsb6\" (UniqueName: \"kubernetes.io/projected/5751e378-39fa-4bf1-98b1-b1242d642c2c-kube-api-access-ktsb6\") pod \"neutron-dhcp-openstack-openstack-cell1-7qr5w\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.284527 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:41:38 crc kubenswrapper[4935]: I1005 09:41:38.803018 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w"] Oct 05 09:41:39 crc kubenswrapper[4935]: I1005 09:41:39.813811 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" event={"ID":"5751e378-39fa-4bf1-98b1-b1242d642c2c","Type":"ContainerStarted","Data":"6915af81e4050346011a289f70f019082cf97fb8447338d23a81bdaa6af60956"} Oct 05 09:41:40 crc kubenswrapper[4935]: I1005 09:41:40.824814 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" event={"ID":"5751e378-39fa-4bf1-98b1-b1242d642c2c","Type":"ContainerStarted","Data":"5fec29c4744d4e068eb0900c65adbfad426c52f3ccd0fce2e93c7e5dba226063"} Oct 05 09:41:40 crc kubenswrapper[4935]: I1005 09:41:40.855441 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" podStartSLOduration=3.347606156 podStartE2EDuration="3.855423641s" podCreationTimestamp="2025-10-05 09:41:37 +0000 UTC" firstStartedPulling="2025-10-05 09:41:39.070300812 +0000 UTC m=+10132.952927272" lastFinishedPulling="2025-10-05 09:41:39.578118287 +0000 UTC m=+10133.460744757" observedRunningTime="2025-10-05 09:41:40.843443856 +0000 UTC m=+10134.726070326" watchObservedRunningTime="2025-10-05 09:41:40.855423641 +0000 UTC m=+10134.738050101" Oct 05 09:42:14 crc kubenswrapper[4935]: I1005 09:42:14.289099 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:42:14 crc kubenswrapper[4935]: I1005 09:42:14.294043 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:42:43 crc kubenswrapper[4935]: I1005 09:42:43.622590 4935 generic.go:334] "Generic (PLEG): container finished" podID="5751e378-39fa-4bf1-98b1-b1242d642c2c" containerID="5fec29c4744d4e068eb0900c65adbfad426c52f3ccd0fce2e93c7e5dba226063" exitCode=0 Oct 05 09:42:43 crc kubenswrapper[4935]: I1005 09:42:43.622662 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" event={"ID":"5751e378-39fa-4bf1-98b1-b1242d642c2c","Type":"ContainerDied","Data":"5fec29c4744d4e068eb0900c65adbfad426c52f3ccd0fce2e93c7e5dba226063"} Oct 05 09:42:44 crc kubenswrapper[4935]: I1005 09:42:44.289258 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:42:44 crc kubenswrapper[4935]: I1005 09:42:44.289330 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.176163 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.287695 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktsb6\" (UniqueName: \"kubernetes.io/projected/5751e378-39fa-4bf1-98b1-b1242d642c2c-kube-api-access-ktsb6\") pod \"5751e378-39fa-4bf1-98b1-b1242d642c2c\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.287781 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-inventory\") pod \"5751e378-39fa-4bf1-98b1-b1242d642c2c\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.287981 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-combined-ca-bundle\") pod \"5751e378-39fa-4bf1-98b1-b1242d642c2c\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.288067 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-agent-neutron-config-0\") pod \"5751e378-39fa-4bf1-98b1-b1242d642c2c\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.288837 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ceph\") pod \"5751e378-39fa-4bf1-98b1-b1242d642c2c\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.288944 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ssh-key\") pod \"5751e378-39fa-4bf1-98b1-b1242d642c2c\" (UID: \"5751e378-39fa-4bf1-98b1-b1242d642c2c\") " Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.294061 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5751e378-39fa-4bf1-98b1-b1242d642c2c-kube-api-access-ktsb6" (OuterVolumeSpecName: "kube-api-access-ktsb6") pod "5751e378-39fa-4bf1-98b1-b1242d642c2c" (UID: "5751e378-39fa-4bf1-98b1-b1242d642c2c"). InnerVolumeSpecName "kube-api-access-ktsb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.294558 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ceph" (OuterVolumeSpecName: "ceph") pod "5751e378-39fa-4bf1-98b1-b1242d642c2c" (UID: "5751e378-39fa-4bf1-98b1-b1242d642c2c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.295957 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "5751e378-39fa-4bf1-98b1-b1242d642c2c" (UID: "5751e378-39fa-4bf1-98b1-b1242d642c2c"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.326251 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "5751e378-39fa-4bf1-98b1-b1242d642c2c" (UID: "5751e378-39fa-4bf1-98b1-b1242d642c2c"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.327716 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-inventory" (OuterVolumeSpecName: "inventory") pod "5751e378-39fa-4bf1-98b1-b1242d642c2c" (UID: "5751e378-39fa-4bf1-98b1-b1242d642c2c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.329442 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5751e378-39fa-4bf1-98b1-b1242d642c2c" (UID: "5751e378-39fa-4bf1-98b1-b1242d642c2c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.391077 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.391111 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktsb6\" (UniqueName: \"kubernetes.io/projected/5751e378-39fa-4bf1-98b1-b1242d642c2c-kube-api-access-ktsb6\") on node \"crc\" DevicePath \"\"" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.391123 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.391132 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.391142 4935 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.391151 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5751e378-39fa-4bf1-98b1-b1242d642c2c-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.655697 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" event={"ID":"5751e378-39fa-4bf1-98b1-b1242d642c2c","Type":"ContainerDied","Data":"6915af81e4050346011a289f70f019082cf97fb8447338d23a81bdaa6af60956"} Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.655766 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-7qr5w" Oct 05 09:42:45 crc kubenswrapper[4935]: I1005 09:42:45.655771 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6915af81e4050346011a289f70f019082cf97fb8447338d23a81bdaa6af60956" Oct 05 09:43:02 crc kubenswrapper[4935]: I1005 09:43:02.007068 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:43:02 crc kubenswrapper[4935]: I1005 09:43:02.007811 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="006b0ab2-2454-4507-92fd-9c314088c339" containerName="nova-cell0-conductor-conductor" containerID="cri-o://aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06" gracePeriod=30 Oct 05 09:43:02 crc kubenswrapper[4935]: I1005 09:43:02.905682 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:43:02 crc kubenswrapper[4935]: I1005 09:43:02.906238 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="7899e5aa-56b1-4016-8e50-97f8ddcfa18f" containerName="nova-cell1-conductor-conductor" containerID="cri-o://d91182927717e44b7a0b70a56d7b9b1546ac136032d8996d1c45922d842f753e" gracePeriod=30 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.079339 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.079567 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d5ff7e17-3100-4f92-a63d-fdd61864a272" containerName="nova-scheduler-scheduler" containerID="cri-o://446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28" gracePeriod=30 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.100633 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.100990 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-log" containerID="cri-o://9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf" gracePeriod=30 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.101505 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-api" containerID="cri-o://8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d" gracePeriod=30 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.115200 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.115431 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-log" containerID="cri-o://27be07cf923f7e51489ec40bf7111a7421289df206b9848709b133d828374621" gracePeriod=30 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.115573 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-metadata" containerID="cri-o://212013d09ca844d3947df7c252d6509ae573f77629ee77ac4068ce10aa8e733b" gracePeriod=30 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.886174 4935 generic.go:334] "Generic (PLEG): container finished" podID="7899e5aa-56b1-4016-8e50-97f8ddcfa18f" containerID="d91182927717e44b7a0b70a56d7b9b1546ac136032d8996d1c45922d842f753e" exitCode=0 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.886247 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7899e5aa-56b1-4016-8e50-97f8ddcfa18f","Type":"ContainerDied","Data":"d91182927717e44b7a0b70a56d7b9b1546ac136032d8996d1c45922d842f753e"} Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.889422 4935 generic.go:334] "Generic (PLEG): container finished" podID="d066b6e6-5017-42da-9868-0d61d229f199" containerID="9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf" exitCode=143 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.889474 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d066b6e6-5017-42da-9868-0d61d229f199","Type":"ContainerDied","Data":"9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf"} Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.891126 4935 generic.go:334] "Generic (PLEG): container finished" podID="bc0befef-da68-4fa7-b273-e030be230f3b" containerID="27be07cf923f7e51489ec40bf7111a7421289df206b9848709b133d828374621" exitCode=143 Oct 05 09:43:03 crc kubenswrapper[4935]: I1005 09:43:03.891151 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc0befef-da68-4fa7-b273-e030be230f3b","Type":"ContainerDied","Data":"27be07cf923f7e51489ec40bf7111a7421289df206b9848709b133d828374621"} Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.271716 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.282016 4935 scope.go:117] "RemoveContainer" containerID="d91182927717e44b7a0b70a56d7b9b1546ac136032d8996d1c45922d842f753e" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.440531 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-combined-ca-bundle\") pod \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.441133 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbm6n\" (UniqueName: \"kubernetes.io/projected/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-kube-api-access-vbm6n\") pod \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.441187 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-config-data\") pod \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\" (UID: \"7899e5aa-56b1-4016-8e50-97f8ddcfa18f\") " Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.469468 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-kube-api-access-vbm6n" (OuterVolumeSpecName: "kube-api-access-vbm6n") pod "7899e5aa-56b1-4016-8e50-97f8ddcfa18f" (UID: "7899e5aa-56b1-4016-8e50-97f8ddcfa18f"). InnerVolumeSpecName "kube-api-access-vbm6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.473802 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7899e5aa-56b1-4016-8e50-97f8ddcfa18f" (UID: "7899e5aa-56b1-4016-8e50-97f8ddcfa18f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.524773 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-config-data" (OuterVolumeSpecName: "config-data") pod "7899e5aa-56b1-4016-8e50-97f8ddcfa18f" (UID: "7899e5aa-56b1-4016-8e50-97f8ddcfa18f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.543266 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.543303 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbm6n\" (UniqueName: \"kubernetes.io/projected/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-kube-api-access-vbm6n\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.543315 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7899e5aa-56b1-4016-8e50-97f8ddcfa18f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.612701 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.746740 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw8c4\" (UniqueName: \"kubernetes.io/projected/006b0ab2-2454-4507-92fd-9c314088c339-kube-api-access-cw8c4\") pod \"006b0ab2-2454-4507-92fd-9c314088c339\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.747018 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-config-data\") pod \"006b0ab2-2454-4507-92fd-9c314088c339\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.747059 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-combined-ca-bundle\") pod \"006b0ab2-2454-4507-92fd-9c314088c339\" (UID: \"006b0ab2-2454-4507-92fd-9c314088c339\") " Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.749618 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/006b0ab2-2454-4507-92fd-9c314088c339-kube-api-access-cw8c4" (OuterVolumeSpecName: "kube-api-access-cw8c4") pod "006b0ab2-2454-4507-92fd-9c314088c339" (UID: "006b0ab2-2454-4507-92fd-9c314088c339"). InnerVolumeSpecName "kube-api-access-cw8c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.775804 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-config-data" (OuterVolumeSpecName: "config-data") pod "006b0ab2-2454-4507-92fd-9c314088c339" (UID: "006b0ab2-2454-4507-92fd-9c314088c339"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.782930 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "006b0ab2-2454-4507-92fd-9c314088c339" (UID: "006b0ab2-2454-4507-92fd-9c314088c339"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.849320 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw8c4\" (UniqueName: \"kubernetes.io/projected/006b0ab2-2454-4507-92fd-9c314088c339-kube-api-access-cw8c4\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.849368 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.849380 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006b0ab2-2454-4507-92fd-9c314088c339-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.902198 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7899e5aa-56b1-4016-8e50-97f8ddcfa18f","Type":"ContainerDied","Data":"d705521a01f1301441db7718da23644c93d16a31cd3130cf6aa6b1910da8fbaf"} Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.904973 4935 generic.go:334] "Generic (PLEG): container finished" podID="006b0ab2-2454-4507-92fd-9c314088c339" containerID="aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06" exitCode=0 Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.905029 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"006b0ab2-2454-4507-92fd-9c314088c339","Type":"ContainerDied","Data":"aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06"} Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.905046 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"006b0ab2-2454-4507-92fd-9c314088c339","Type":"ContainerDied","Data":"9bb6e092776f8d3cb38e9015c6cd0f5cb812b531d1609c7bd027e46a361b0f26"} Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.905061 4935 scope.go:117] "RemoveContainer" containerID="aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.905057 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.910696 4935 generic.go:334] "Generic (PLEG): container finished" podID="d5ff7e17-3100-4f92-a63d-fdd61864a272" containerID="446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28" exitCode=0 Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.910770 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.910796 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d5ff7e17-3100-4f92-a63d-fdd61864a272","Type":"ContainerDied","Data":"446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28"} Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.951327 4935 scope.go:117] "RemoveContainer" containerID="aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06" Oct 05 09:43:04 crc kubenswrapper[4935]: E1005 09:43:04.952541 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06\": container with ID starting with aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06 not found: ID does not exist" containerID="aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.952686 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06"} err="failed to get container status \"aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06\": rpc error: code = NotFound desc = could not find container \"aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06\": container with ID starting with aad7e543dc4946f5ba9fb6e787b480ab68182ae5904e31326aa71a5064474e06 not found: ID does not exist" Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.972815 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:43:04 crc kubenswrapper[4935]: I1005 09:43:04.990439 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.005962 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.016126 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:43:05 crc kubenswrapper[4935]: E1005 09:43:05.024084 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28 is running failed: container process not found" containerID="446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 09:43:05 crc kubenswrapper[4935]: E1005 09:43:05.024386 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28 is running failed: container process not found" containerID="446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.024460 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:43:05 crc kubenswrapper[4935]: E1005 09:43:05.024904 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5751e378-39fa-4bf1-98b1-b1242d642c2c" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.024915 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5751e378-39fa-4bf1-98b1-b1242d642c2c" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 05 09:43:05 crc kubenswrapper[4935]: E1005 09:43:05.024930 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7899e5aa-56b1-4016-8e50-97f8ddcfa18f" containerName="nova-cell1-conductor-conductor" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.024937 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7899e5aa-56b1-4016-8e50-97f8ddcfa18f" containerName="nova-cell1-conductor-conductor" Oct 05 09:43:05 crc kubenswrapper[4935]: E1005 09:43:05.024963 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006b0ab2-2454-4507-92fd-9c314088c339" containerName="nova-cell0-conductor-conductor" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.024969 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="006b0ab2-2454-4507-92fd-9c314088c339" containerName="nova-cell0-conductor-conductor" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.025163 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7899e5aa-56b1-4016-8e50-97f8ddcfa18f" containerName="nova-cell1-conductor-conductor" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.025180 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="006b0ab2-2454-4507-92fd-9c314088c339" containerName="nova-cell0-conductor-conductor" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.025197 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5751e378-39fa-4bf1-98b1-b1242d642c2c" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 05 09:43:05 crc kubenswrapper[4935]: E1005 09:43:05.025237 4935 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28 is running failed: container process not found" containerID="446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 09:43:05 crc kubenswrapper[4935]: E1005 09:43:05.025288 4935 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d5ff7e17-3100-4f92-a63d-fdd61864a272" containerName="nova-scheduler-scheduler" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.026108 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.028245 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.036368 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.047281 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.048798 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.059424 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.062497 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.156744 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j698t\" (UniqueName: \"kubernetes.io/projected/e91bdcc1-916c-432a-8df0-ee70b173a847-kube-api-access-j698t\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.158864 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e91bdcc1-916c-432a-8df0-ee70b173a847-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.159076 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e91bdcc1-916c-432a-8df0-ee70b173a847-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.159212 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.159239 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.159286 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmw42\" (UniqueName: \"kubernetes.io/projected/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-kube-api-access-rmw42\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.175488 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.260674 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j698t\" (UniqueName: \"kubernetes.io/projected/e91bdcc1-916c-432a-8df0-ee70b173a847-kube-api-access-j698t\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.260733 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e91bdcc1-916c-432a-8df0-ee70b173a847-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.260797 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e91bdcc1-916c-432a-8df0-ee70b173a847-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.260861 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.260906 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.260951 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmw42\" (UniqueName: \"kubernetes.io/projected/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-kube-api-access-rmw42\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.266424 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e91bdcc1-916c-432a-8df0-ee70b173a847-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.266644 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.267501 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e91bdcc1-916c-432a-8df0-ee70b173a847-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.268334 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.276862 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmw42\" (UniqueName: \"kubernetes.io/projected/7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee-kube-api-access-rmw42\") pod \"nova-cell1-conductor-0\" (UID: \"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.277852 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j698t\" (UniqueName: \"kubernetes.io/projected/e91bdcc1-916c-432a-8df0-ee70b173a847-kube-api-access-j698t\") pod \"nova-cell0-conductor-0\" (UID: \"e91bdcc1-916c-432a-8df0-ee70b173a847\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.361953 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-combined-ca-bundle\") pod \"d5ff7e17-3100-4f92-a63d-fdd61864a272\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.362700 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvd2k\" (UniqueName: \"kubernetes.io/projected/d5ff7e17-3100-4f92-a63d-fdd61864a272-kube-api-access-vvd2k\") pod \"d5ff7e17-3100-4f92-a63d-fdd61864a272\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.362882 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-config-data\") pod \"d5ff7e17-3100-4f92-a63d-fdd61864a272\" (UID: \"d5ff7e17-3100-4f92-a63d-fdd61864a272\") " Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.368562 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.368993 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5ff7e17-3100-4f92-a63d-fdd61864a272-kube-api-access-vvd2k" (OuterVolumeSpecName: "kube-api-access-vvd2k") pod "d5ff7e17-3100-4f92-a63d-fdd61864a272" (UID: "d5ff7e17-3100-4f92-a63d-fdd61864a272"). InnerVolumeSpecName "kube-api-access-vvd2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.389203 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.395031 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-config-data" (OuterVolumeSpecName: "config-data") pod "d5ff7e17-3100-4f92-a63d-fdd61864a272" (UID: "d5ff7e17-3100-4f92-a63d-fdd61864a272"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.408746 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5ff7e17-3100-4f92-a63d-fdd61864a272" (UID: "d5ff7e17-3100-4f92-a63d-fdd61864a272"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.465625 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvd2k\" (UniqueName: \"kubernetes.io/projected/d5ff7e17-3100-4f92-a63d-fdd61864a272-kube-api-access-vvd2k\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.465666 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.465678 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5ff7e17-3100-4f92-a63d-fdd61864a272-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.895852 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.927609 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d5ff7e17-3100-4f92-a63d-fdd61864a272","Type":"ContainerDied","Data":"2755422a86994a32e0eee1a1ac2e4314171ac81769b54ecc04af30f7812f8ef7"} Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.927674 4935 scope.go:117] "RemoveContainer" containerID="446c5848bdb31eb508e464515e61ebae4597ffb5a210ceae7c2cb9c1681cce28" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.927922 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.930012 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee","Type":"ContainerStarted","Data":"265749af417be7a1fc6598efca9ce2de84c174944f0a4084e294f87a0c679ad9"} Oct 05 09:43:05 crc kubenswrapper[4935]: W1005 09:43:05.988963 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode91bdcc1_916c_432a_8df0_ee70b173a847.slice/crio-e025000bc2670a155d2203fb80e111eda3b917bfa3c80c472b8c27d77d162dea WatchSource:0}: Error finding container e025000bc2670a155d2203fb80e111eda3b917bfa3c80c472b8c27d77d162dea: Status 404 returned error can't find the container with id e025000bc2670a155d2203fb80e111eda3b917bfa3c80c472b8c27d77d162dea Oct 05 09:43:05 crc kubenswrapper[4935]: I1005 09:43:05.993602 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.119001 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.130344 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.147131 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:43:06 crc kubenswrapper[4935]: E1005 09:43:06.147714 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5ff7e17-3100-4f92-a63d-fdd61864a272" containerName="nova-scheduler-scheduler" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.147736 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5ff7e17-3100-4f92-a63d-fdd61864a272" containerName="nova-scheduler-scheduler" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.148083 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5ff7e17-3100-4f92-a63d-fdd61864a272" containerName="nova-scheduler-scheduler" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.149180 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.151238 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.172274 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.185448 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdb2053-f3c5-4665-a7c7-80f708039ae1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.185502 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgtp7\" (UniqueName: \"kubernetes.io/projected/afdb2053-f3c5-4665-a7c7-80f708039ae1-kube-api-access-zgtp7\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.185612 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdb2053-f3c5-4665-a7c7-80f708039ae1-config-data\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.287496 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdb2053-f3c5-4665-a7c7-80f708039ae1-config-data\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.287668 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdb2053-f3c5-4665-a7c7-80f708039ae1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.287710 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgtp7\" (UniqueName: \"kubernetes.io/projected/afdb2053-f3c5-4665-a7c7-80f708039ae1-kube-api-access-zgtp7\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.296015 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdb2053-f3c5-4665-a7c7-80f708039ae1-config-data\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.299158 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdb2053-f3c5-4665-a7c7-80f708039ae1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.310467 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgtp7\" (UniqueName: \"kubernetes.io/projected/afdb2053-f3c5-4665-a7c7-80f708039ae1-kube-api-access-zgtp7\") pod \"nova-scheduler-0\" (UID: \"afdb2053-f3c5-4665-a7c7-80f708039ae1\") " pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.472655 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.812338 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="006b0ab2-2454-4507-92fd-9c314088c339" path="/var/lib/kubelet/pods/006b0ab2-2454-4507-92fd-9c314088c339/volumes" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.813598 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7899e5aa-56b1-4016-8e50-97f8ddcfa18f" path="/var/lib/kubelet/pods/7899e5aa-56b1-4016-8e50-97f8ddcfa18f/volumes" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.814172 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5ff7e17-3100-4f92-a63d-fdd61864a272" path="/var/lib/kubelet/pods/d5ff7e17-3100-4f92-a63d-fdd61864a272/volumes" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.816832 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.948377 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee","Type":"ContainerStarted","Data":"89d1dcaae70433b30a6845b5cd441b854de441c49e9bfcaf1424c85216c63d37"} Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.950176 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.953304 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e91bdcc1-916c-432a-8df0-ee70b173a847","Type":"ContainerStarted","Data":"750442e153283fc890c6703ba30efc5a20e72990eef4dc6a91f672525a871634"} Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.953403 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e91bdcc1-916c-432a-8df0-ee70b173a847","Type":"ContainerStarted","Data":"e025000bc2670a155d2203fb80e111eda3b917bfa3c80c472b8c27d77d162dea"} Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.955322 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:06 crc kubenswrapper[4935]: W1005 09:43:06.982002 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafdb2053_f3c5_4665_a7c7_80f708039ae1.slice/crio-d6dfb07d76e55c20eb4286f54b237e30f4475465de379df76ba01ffd9061a165 WatchSource:0}: Error finding container d6dfb07d76e55c20eb4286f54b237e30f4475465de379df76ba01ffd9061a165: Status 404 returned error can't find the container with id d6dfb07d76e55c20eb4286f54b237e30f4475465de379df76ba01ffd9061a165 Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.982160 4935 generic.go:334] "Generic (PLEG): container finished" podID="d066b6e6-5017-42da-9868-0d61d229f199" containerID="8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d" exitCode=0 Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.982228 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d066b6e6-5017-42da-9868-0d61d229f199","Type":"ContainerDied","Data":"8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d"} Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.982260 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d066b6e6-5017-42da-9868-0d61d229f199","Type":"ContainerDied","Data":"be36e01b91f6ae31c3dd4f6162e509a2da661bf0725d08de3b534d33228bbe54"} Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.982277 4935 scope.go:117] "RemoveContainer" containerID="8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.982305 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 09:43:06 crc kubenswrapper[4935]: I1005 09:43:06.994970 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.013602 4935 generic.go:334] "Generic (PLEG): container finished" podID="bc0befef-da68-4fa7-b273-e030be230f3b" containerID="212013d09ca844d3947df7c252d6509ae573f77629ee77ac4068ce10aa8e733b" exitCode=0 Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.013649 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d066b6e6-5017-42da-9868-0d61d229f199-logs\") pod \"d066b6e6-5017-42da-9868-0d61d229f199\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.013721 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-config-data\") pod \"d066b6e6-5017-42da-9868-0d61d229f199\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.013755 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p86wl\" (UniqueName: \"kubernetes.io/projected/d066b6e6-5017-42da-9868-0d61d229f199-kube-api-access-p86wl\") pod \"d066b6e6-5017-42da-9868-0d61d229f199\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.013835 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-combined-ca-bundle\") pod \"d066b6e6-5017-42da-9868-0d61d229f199\" (UID: \"d066b6e6-5017-42da-9868-0d61d229f199\") " Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.014385 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d066b6e6-5017-42da-9868-0d61d229f199-logs" (OuterVolumeSpecName: "logs") pod "d066b6e6-5017-42da-9868-0d61d229f199" (UID: "d066b6e6-5017-42da-9868-0d61d229f199"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.013651 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc0befef-da68-4fa7-b273-e030be230f3b","Type":"ContainerDied","Data":"212013d09ca844d3947df7c252d6509ae573f77629ee77ac4068ce10aa8e733b"} Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.032062 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d066b6e6-5017-42da-9868-0d61d229f199-kube-api-access-p86wl" (OuterVolumeSpecName: "kube-api-access-p86wl") pod "d066b6e6-5017-42da-9868-0d61d229f199" (UID: "d066b6e6-5017-42da-9868-0d61d229f199"). InnerVolumeSpecName "kube-api-access-p86wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.032663 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.032644228 podStartE2EDuration="3.032644228s" podCreationTimestamp="2025-10-05 09:43:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:43:06.967656266 +0000 UTC m=+10220.850282726" watchObservedRunningTime="2025-10-05 09:43:07.032644228 +0000 UTC m=+10220.915270688" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.049666 4935 scope.go:117] "RemoveContainer" containerID="9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.068441 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-config-data" (OuterVolumeSpecName: "config-data") pod "d066b6e6-5017-42da-9868-0d61d229f199" (UID: "d066b6e6-5017-42da-9868-0d61d229f199"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.077045 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d066b6e6-5017-42da-9868-0d61d229f199" (UID: "d066b6e6-5017-42da-9868-0d61d229f199"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.084260 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.084219364 podStartE2EDuration="3.084219364s" podCreationTimestamp="2025-10-05 09:43:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:43:06.98931234 +0000 UTC m=+10220.871938800" watchObservedRunningTime="2025-10-05 09:43:07.084219364 +0000 UTC m=+10220.966845824" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.117347 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d066b6e6-5017-42da-9868-0d61d229f199-logs\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.117387 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.117400 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p86wl\" (UniqueName: \"kubernetes.io/projected/d066b6e6-5017-42da-9868-0d61d229f199-kube-api-access-p86wl\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.117411 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d066b6e6-5017-42da-9868-0d61d229f199-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.139168 4935 scope.go:117] "RemoveContainer" containerID="8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d" Oct 05 09:43:07 crc kubenswrapper[4935]: E1005 09:43:07.143161 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d\": container with ID starting with 8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d not found: ID does not exist" containerID="8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.143192 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d"} err="failed to get container status \"8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d\": rpc error: code = NotFound desc = could not find container \"8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d\": container with ID starting with 8e1a3f2f457a87f0c5579ed470e31a4dceee34402270914297d3e35f6b24e96d not found: ID does not exist" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.143215 4935 scope.go:117] "RemoveContainer" containerID="9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf" Oct 05 09:43:07 crc kubenswrapper[4935]: E1005 09:43:07.144189 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf\": container with ID starting with 9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf not found: ID does not exist" containerID="9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.144255 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf"} err="failed to get container status \"9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf\": rpc error: code = NotFound desc = could not find container \"9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf\": container with ID starting with 9a0cac7ec9d87a7b2986581923e27f759b56a20f56a129481cf80f9cd70beeaf not found: ID does not exist" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.157800 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.320631 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-combined-ca-bundle\") pod \"bc0befef-da68-4fa7-b273-e030be230f3b\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.321112 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blf55\" (UniqueName: \"kubernetes.io/projected/bc0befef-da68-4fa7-b273-e030be230f3b-kube-api-access-blf55\") pod \"bc0befef-da68-4fa7-b273-e030be230f3b\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.321240 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc0befef-da68-4fa7-b273-e030be230f3b-logs\") pod \"bc0befef-da68-4fa7-b273-e030be230f3b\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.321317 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-config-data\") pod \"bc0befef-da68-4fa7-b273-e030be230f3b\" (UID: \"bc0befef-da68-4fa7-b273-e030be230f3b\") " Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.332280 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc0befef-da68-4fa7-b273-e030be230f3b-logs" (OuterVolumeSpecName: "logs") pod "bc0befef-da68-4fa7-b273-e030be230f3b" (UID: "bc0befef-da68-4fa7-b273-e030be230f3b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.333635 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0befef-da68-4fa7-b273-e030be230f3b-kube-api-access-blf55" (OuterVolumeSpecName: "kube-api-access-blf55") pod "bc0befef-da68-4fa7-b273-e030be230f3b" (UID: "bc0befef-da68-4fa7-b273-e030be230f3b"). InnerVolumeSpecName "kube-api-access-blf55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.351906 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.359687 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc0befef-da68-4fa7-b273-e030be230f3b" (UID: "bc0befef-da68-4fa7-b273-e030be230f3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.363464 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.379687 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 09:43:07 crc kubenswrapper[4935]: E1005 09:43:07.380317 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-log" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.380329 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-log" Oct 05 09:43:07 crc kubenswrapper[4935]: E1005 09:43:07.380350 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-api" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.380355 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-api" Oct 05 09:43:07 crc kubenswrapper[4935]: E1005 09:43:07.380380 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-log" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.380388 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-log" Oct 05 09:43:07 crc kubenswrapper[4935]: E1005 09:43:07.380412 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-metadata" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.380420 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-metadata" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.380595 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-log" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.380613 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-metadata" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.380627 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" containerName="nova-metadata-log" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.380640 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d066b6e6-5017-42da-9868-0d61d229f199" containerName="nova-api-api" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.381767 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.386213 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.400344 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.414567 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-config-data" (OuterVolumeSpecName: "config-data") pod "bc0befef-da68-4fa7-b273-e030be230f3b" (UID: "bc0befef-da68-4fa7-b273-e030be230f3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.423165 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.423409 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc0befef-da68-4fa7-b273-e030be230f3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.423474 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blf55\" (UniqueName: \"kubernetes.io/projected/bc0befef-da68-4fa7-b273-e030be230f3b-kube-api-access-blf55\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.423544 4935 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc0befef-da68-4fa7-b273-e030be230f3b-logs\") on node \"crc\" DevicePath \"\"" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.525977 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c69aa404-fec4-4ba7-bafe-30c254ac2d38-logs\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.526220 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c69aa404-fec4-4ba7-bafe-30c254ac2d38-config-data\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.526262 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q74g8\" (UniqueName: \"kubernetes.io/projected/c69aa404-fec4-4ba7-bafe-30c254ac2d38-kube-api-access-q74g8\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.526401 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c69aa404-fec4-4ba7-bafe-30c254ac2d38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.630193 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c69aa404-fec4-4ba7-bafe-30c254ac2d38-logs\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.630323 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c69aa404-fec4-4ba7-bafe-30c254ac2d38-config-data\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.630368 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q74g8\" (UniqueName: \"kubernetes.io/projected/c69aa404-fec4-4ba7-bafe-30c254ac2d38-kube-api-access-q74g8\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.630407 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c69aa404-fec4-4ba7-bafe-30c254ac2d38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.630716 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c69aa404-fec4-4ba7-bafe-30c254ac2d38-logs\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.641081 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c69aa404-fec4-4ba7-bafe-30c254ac2d38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.644707 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c69aa404-fec4-4ba7-bafe-30c254ac2d38-config-data\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.659070 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q74g8\" (UniqueName: \"kubernetes.io/projected/c69aa404-fec4-4ba7-bafe-30c254ac2d38-kube-api-access-q74g8\") pod \"nova-api-0\" (UID: \"c69aa404-fec4-4ba7-bafe-30c254ac2d38\") " pod="openstack/nova-api-0" Oct 05 09:43:07 crc kubenswrapper[4935]: I1005 09:43:07.706725 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.025319 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"afdb2053-f3c5-4665-a7c7-80f708039ae1","Type":"ContainerStarted","Data":"f130251349f128d8a46655591754cc81f48867d5dda068e255271f6066f10934"} Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.025370 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"afdb2053-f3c5-4665-a7c7-80f708039ae1","Type":"ContainerStarted","Data":"d6dfb07d76e55c20eb4286f54b237e30f4475465de379df76ba01ffd9061a165"} Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.032807 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc0befef-da68-4fa7-b273-e030be230f3b","Type":"ContainerDied","Data":"05599728c663204ab9d7f2bc3a07a6c97a11623a0d83265ba865b43ac30f3e58"} Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.032907 4935 scope.go:117] "RemoveContainer" containerID="212013d09ca844d3947df7c252d6509ae573f77629ee77ac4068ce10aa8e733b" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.032993 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.051298 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.051277103 podStartE2EDuration="2.051277103s" podCreationTimestamp="2025-10-05 09:43:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:43:08.042900321 +0000 UTC m=+10221.925526781" watchObservedRunningTime="2025-10-05 09:43:08.051277103 +0000 UTC m=+10221.933903563" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.069182 4935 scope.go:117] "RemoveContainer" containerID="27be07cf923f7e51489ec40bf7111a7421289df206b9848709b133d828374621" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.089710 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.116179 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.151372 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.164296 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.167228 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.168507 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.188469 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.274273 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppqcw\" (UniqueName: \"kubernetes.io/projected/e6cc20f0-262c-4d74-b562-72e20576637f-kube-api-access-ppqcw\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.274674 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6cc20f0-262c-4d74-b562-72e20576637f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.274844 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6cc20f0-262c-4d74-b562-72e20576637f-logs\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.274951 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6cc20f0-262c-4d74-b562-72e20576637f-config-data\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.376698 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppqcw\" (UniqueName: \"kubernetes.io/projected/e6cc20f0-262c-4d74-b562-72e20576637f-kube-api-access-ppqcw\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.376818 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6cc20f0-262c-4d74-b562-72e20576637f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.376911 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6cc20f0-262c-4d74-b562-72e20576637f-logs\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.376931 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6cc20f0-262c-4d74-b562-72e20576637f-config-data\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.377692 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6cc20f0-262c-4d74-b562-72e20576637f-logs\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.380810 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6cc20f0-262c-4d74-b562-72e20576637f-config-data\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.381193 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6cc20f0-262c-4d74-b562-72e20576637f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.392438 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppqcw\" (UniqueName: \"kubernetes.io/projected/e6cc20f0-262c-4d74-b562-72e20576637f-kube-api-access-ppqcw\") pod \"nova-metadata-0\" (UID: \"e6cc20f0-262c-4d74-b562-72e20576637f\") " pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.531114 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.797456 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc0befef-da68-4fa7-b273-e030be230f3b" path="/var/lib/kubelet/pods/bc0befef-da68-4fa7-b273-e030be230f3b/volumes" Oct 05 09:43:08 crc kubenswrapper[4935]: I1005 09:43:08.798577 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d066b6e6-5017-42da-9868-0d61d229f199" path="/var/lib/kubelet/pods/d066b6e6-5017-42da-9868-0d61d229f199/volumes" Oct 05 09:43:09 crc kubenswrapper[4935]: I1005 09:43:09.003384 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:43:09 crc kubenswrapper[4935]: I1005 09:43:09.051522 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c69aa404-fec4-4ba7-bafe-30c254ac2d38","Type":"ContainerStarted","Data":"5fefcc64b5d0114bee026b346b05967f199a682d4d123d86ef169f3743527cc7"} Oct 05 09:43:09 crc kubenswrapper[4935]: I1005 09:43:09.051836 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c69aa404-fec4-4ba7-bafe-30c254ac2d38","Type":"ContainerStarted","Data":"3d94ba6c07406850894cd4c78a8ad50e1c58a18eff043e27d470103e71ce4d18"} Oct 05 09:43:09 crc kubenswrapper[4935]: I1005 09:43:09.051848 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c69aa404-fec4-4ba7-bafe-30c254ac2d38","Type":"ContainerStarted","Data":"66380e90df50219510b20640d0ac0e5146c8fa0dd4a3f4b3822204ef57000b3a"} Oct 05 09:43:09 crc kubenswrapper[4935]: I1005 09:43:09.057052 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6cc20f0-262c-4d74-b562-72e20576637f","Type":"ContainerStarted","Data":"15751568e51e0755b4b59002e2bf4ea508c0fae55a74748fee82c665f7be1a76"} Oct 05 09:43:09 crc kubenswrapper[4935]: I1005 09:43:09.073373 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.073349309 podStartE2EDuration="2.073349309s" podCreationTimestamp="2025-10-05 09:43:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:43:09.069541218 +0000 UTC m=+10222.952167668" watchObservedRunningTime="2025-10-05 09:43:09.073349309 +0000 UTC m=+10222.955975799" Oct 05 09:43:10 crc kubenswrapper[4935]: I1005 09:43:10.066785 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6cc20f0-262c-4d74-b562-72e20576637f","Type":"ContainerStarted","Data":"01b9fbc997f4b48c3fe42ae36979a14b3216a1259ddc599d3be7d0ffb1d302dc"} Oct 05 09:43:10 crc kubenswrapper[4935]: I1005 09:43:10.067089 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6cc20f0-262c-4d74-b562-72e20576637f","Type":"ContainerStarted","Data":"9e73e1afc4eee0c747b32a180c12cfe18ddc81803939a4959e717429a6357a97"} Oct 05 09:43:10 crc kubenswrapper[4935]: I1005 09:43:10.089475 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.089453467 podStartE2EDuration="2.089453467s" podCreationTimestamp="2025-10-05 09:43:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:43:10.081756033 +0000 UTC m=+10223.964382493" watchObservedRunningTime="2025-10-05 09:43:10.089453467 +0000 UTC m=+10223.972079927" Oct 05 09:43:11 crc kubenswrapper[4935]: I1005 09:43:11.474096 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 09:43:13 crc kubenswrapper[4935]: I1005 09:43:13.531500 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 09:43:13 crc kubenswrapper[4935]: I1005 09:43:13.531961 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 09:43:14 crc kubenswrapper[4935]: I1005 09:43:14.289163 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:43:14 crc kubenswrapper[4935]: I1005 09:43:14.289616 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:43:14 crc kubenswrapper[4935]: I1005 09:43:14.289696 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:43:14 crc kubenswrapper[4935]: I1005 09:43:14.290593 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c31b5e83aad13ab5aad5147c3c65b8e692b7949147dda7547b1fcc298444ac8b"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:43:14 crc kubenswrapper[4935]: I1005 09:43:14.290754 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://c31b5e83aad13ab5aad5147c3c65b8e692b7949147dda7547b1fcc298444ac8b" gracePeriod=600 Oct 05 09:43:15 crc kubenswrapper[4935]: I1005 09:43:15.136804 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="c31b5e83aad13ab5aad5147c3c65b8e692b7949147dda7547b1fcc298444ac8b" exitCode=0 Oct 05 09:43:15 crc kubenswrapper[4935]: I1005 09:43:15.136883 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"c31b5e83aad13ab5aad5147c3c65b8e692b7949147dda7547b1fcc298444ac8b"} Oct 05 09:43:15 crc kubenswrapper[4935]: I1005 09:43:15.137735 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7"} Oct 05 09:43:15 crc kubenswrapper[4935]: I1005 09:43:15.137782 4935 scope.go:117] "RemoveContainer" containerID="b2cb0a86669b09777450af3062f0c19f6976b755e13244f52557c8459827f3c9" Oct 05 09:43:15 crc kubenswrapper[4935]: I1005 09:43:15.491740 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 05 09:43:15 crc kubenswrapper[4935]: I1005 09:43:15.503715 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 05 09:43:16 crc kubenswrapper[4935]: I1005 09:43:16.473886 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 09:43:16 crc kubenswrapper[4935]: I1005 09:43:16.516406 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 09:43:17 crc kubenswrapper[4935]: I1005 09:43:17.190230 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 09:43:17 crc kubenswrapper[4935]: I1005 09:43:17.708111 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 09:43:17 crc kubenswrapper[4935]: I1005 09:43:17.708182 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 09:43:18 crc kubenswrapper[4935]: I1005 09:43:18.532047 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 09:43:18 crc kubenswrapper[4935]: I1005 09:43:18.532103 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 09:43:18 crc kubenswrapper[4935]: I1005 09:43:18.792124 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c69aa404-fec4-4ba7-bafe-30c254ac2d38" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 09:43:18 crc kubenswrapper[4935]: I1005 09:43:18.792135 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c69aa404-fec4-4ba7-bafe-30c254ac2d38" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 09:43:19 crc kubenswrapper[4935]: I1005 09:43:19.615034 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e6cc20f0-262c-4d74-b562-72e20576637f" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 09:43:19 crc kubenswrapper[4935]: I1005 09:43:19.615034 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e6cc20f0-262c-4d74-b562-72e20576637f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 09:43:27 crc kubenswrapper[4935]: I1005 09:43:27.711055 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 09:43:27 crc kubenswrapper[4935]: I1005 09:43:27.711739 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 09:43:27 crc kubenswrapper[4935]: I1005 09:43:27.712400 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 09:43:27 crc kubenswrapper[4935]: I1005 09:43:27.712429 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 09:43:27 crc kubenswrapper[4935]: I1005 09:43:27.715450 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 09:43:27 crc kubenswrapper[4935]: I1005 09:43:27.716285 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 09:43:28 crc kubenswrapper[4935]: I1005 09:43:28.534453 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 09:43:28 crc kubenswrapper[4935]: I1005 09:43:28.534996 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 09:43:28 crc kubenswrapper[4935]: I1005 09:43:28.538244 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 09:43:29 crc kubenswrapper[4935]: I1005 09:43:29.306114 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.494018 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4"] Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.496372 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.501652 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.502633 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.502927 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.503073 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.503242 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.503380 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-t5sss" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.503810 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.511874 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4"] Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.607969 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.608159 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfkp8\" (UniqueName: \"kubernetes.io/projected/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-kube-api-access-pfkp8\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.608312 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.608407 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.608605 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.608693 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.608737 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.608834 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.608912 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.609037 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.609065 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710401 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710455 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710516 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710542 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710562 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710586 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710781 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710830 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710850 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710881 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.710952 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfkp8\" (UniqueName: \"kubernetes.io/projected/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-kube-api-access-pfkp8\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.712026 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.712084 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.720645 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.720692 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.724079 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.724355 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.724428 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.724444 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.724566 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.724756 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.733023 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfkp8\" (UniqueName: \"kubernetes.io/projected/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-kube-api-access-pfkp8\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:30 crc kubenswrapper[4935]: I1005 09:43:30.818506 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:43:31 crc kubenswrapper[4935]: I1005 09:43:31.365869 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4"] Oct 05 09:43:31 crc kubenswrapper[4935]: W1005 09:43:31.368945 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a8cc4c7_3c2c_4f19_9224_8d06432ca678.slice/crio-89e5e9ca7b471c27a9a2a834a98b851cdf601dc4d7ac7ccd0f05f98d4d2c4767 WatchSource:0}: Error finding container 89e5e9ca7b471c27a9a2a834a98b851cdf601dc4d7ac7ccd0f05f98d4d2c4767: Status 404 returned error can't find the container with id 89e5e9ca7b471c27a9a2a834a98b851cdf601dc4d7ac7ccd0f05f98d4d2c4767 Oct 05 09:43:32 crc kubenswrapper[4935]: I1005 09:43:32.345379 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" event={"ID":"8a8cc4c7-3c2c-4f19-9224-8d06432ca678","Type":"ContainerStarted","Data":"ecf203170c5eb69fbd47d67efa205ac6d0bf2c5f0988f21c49f726e85369f72b"} Oct 05 09:43:32 crc kubenswrapper[4935]: I1005 09:43:32.346035 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" event={"ID":"8a8cc4c7-3c2c-4f19-9224-8d06432ca678","Type":"ContainerStarted","Data":"89e5e9ca7b471c27a9a2a834a98b851cdf601dc4d7ac7ccd0f05f98d4d2c4767"} Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.162803 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" podStartSLOduration=89.646690708 podStartE2EDuration="1m30.16277854s" podCreationTimestamp="2025-10-05 09:43:30 +0000 UTC" firstStartedPulling="2025-10-05 09:43:31.372627262 +0000 UTC m=+10245.255253742" lastFinishedPulling="2025-10-05 09:43:31.888715114 +0000 UTC m=+10245.771341574" observedRunningTime="2025-10-05 09:43:32.371992157 +0000 UTC m=+10246.254618627" watchObservedRunningTime="2025-10-05 09:45:00.16277854 +0000 UTC m=+10334.045405000" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.168814 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq"] Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.171051 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.175050 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.177137 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.184745 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq"] Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.282294 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvtq8\" (UniqueName: \"kubernetes.io/projected/053eb908-340c-49d3-9ce3-0b126f4f3d44-kube-api-access-hvtq8\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.282432 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/053eb908-340c-49d3-9ce3-0b126f4f3d44-config-volume\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.282512 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/053eb908-340c-49d3-9ce3-0b126f4f3d44-secret-volume\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.384863 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/053eb908-340c-49d3-9ce3-0b126f4f3d44-secret-volume\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.385081 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvtq8\" (UniqueName: \"kubernetes.io/projected/053eb908-340c-49d3-9ce3-0b126f4f3d44-kube-api-access-hvtq8\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.385145 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/053eb908-340c-49d3-9ce3-0b126f4f3d44-config-volume\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.386410 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/053eb908-340c-49d3-9ce3-0b126f4f3d44-config-volume\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.400078 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/053eb908-340c-49d3-9ce3-0b126f4f3d44-secret-volume\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.404882 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvtq8\" (UniqueName: \"kubernetes.io/projected/053eb908-340c-49d3-9ce3-0b126f4f3d44-kube-api-access-hvtq8\") pod \"collect-profiles-29327625-46bsq\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:00 crc kubenswrapper[4935]: I1005 09:45:00.500868 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:01 crc kubenswrapper[4935]: I1005 09:45:00.984090 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq"] Oct 05 09:45:01 crc kubenswrapper[4935]: I1005 09:45:01.430387 4935 generic.go:334] "Generic (PLEG): container finished" podID="053eb908-340c-49d3-9ce3-0b126f4f3d44" containerID="afd352ee33ce9b0cd1f0b705e78d166423b7730ac592ca91ed33d77052c5335b" exitCode=0 Oct 05 09:45:01 crc kubenswrapper[4935]: I1005 09:45:01.430711 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" event={"ID":"053eb908-340c-49d3-9ce3-0b126f4f3d44","Type":"ContainerDied","Data":"afd352ee33ce9b0cd1f0b705e78d166423b7730ac592ca91ed33d77052c5335b"} Oct 05 09:45:01 crc kubenswrapper[4935]: I1005 09:45:01.430747 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" event={"ID":"053eb908-340c-49d3-9ce3-0b126f4f3d44","Type":"ContainerStarted","Data":"42edf8d3cee07146a2f2932083190709511868889085c1349305daa5945ef09e"} Oct 05 09:45:02 crc kubenswrapper[4935]: I1005 09:45:02.825467 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:02 crc kubenswrapper[4935]: I1005 09:45:02.950733 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/053eb908-340c-49d3-9ce3-0b126f4f3d44-secret-volume\") pod \"053eb908-340c-49d3-9ce3-0b126f4f3d44\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " Oct 05 09:45:02 crc kubenswrapper[4935]: I1005 09:45:02.951061 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvtq8\" (UniqueName: \"kubernetes.io/projected/053eb908-340c-49d3-9ce3-0b126f4f3d44-kube-api-access-hvtq8\") pod \"053eb908-340c-49d3-9ce3-0b126f4f3d44\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " Oct 05 09:45:02 crc kubenswrapper[4935]: I1005 09:45:02.951220 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/053eb908-340c-49d3-9ce3-0b126f4f3d44-config-volume\") pod \"053eb908-340c-49d3-9ce3-0b126f4f3d44\" (UID: \"053eb908-340c-49d3-9ce3-0b126f4f3d44\") " Oct 05 09:45:02 crc kubenswrapper[4935]: I1005 09:45:02.951673 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/053eb908-340c-49d3-9ce3-0b126f4f3d44-config-volume" (OuterVolumeSpecName: "config-volume") pod "053eb908-340c-49d3-9ce3-0b126f4f3d44" (UID: "053eb908-340c-49d3-9ce3-0b126f4f3d44"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:45:02 crc kubenswrapper[4935]: I1005 09:45:02.952074 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/053eb908-340c-49d3-9ce3-0b126f4f3d44-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:02 crc kubenswrapper[4935]: I1005 09:45:02.955984 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053eb908-340c-49d3-9ce3-0b126f4f3d44-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "053eb908-340c-49d3-9ce3-0b126f4f3d44" (UID: "053eb908-340c-49d3-9ce3-0b126f4f3d44"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:45:02 crc kubenswrapper[4935]: I1005 09:45:02.956717 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053eb908-340c-49d3-9ce3-0b126f4f3d44-kube-api-access-hvtq8" (OuterVolumeSpecName: "kube-api-access-hvtq8") pod "053eb908-340c-49d3-9ce3-0b126f4f3d44" (UID: "053eb908-340c-49d3-9ce3-0b126f4f3d44"). InnerVolumeSpecName "kube-api-access-hvtq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:45:03 crc kubenswrapper[4935]: I1005 09:45:03.054529 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/053eb908-340c-49d3-9ce3-0b126f4f3d44-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:03 crc kubenswrapper[4935]: I1005 09:45:03.054584 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvtq8\" (UniqueName: \"kubernetes.io/projected/053eb908-340c-49d3-9ce3-0b126f4f3d44-kube-api-access-hvtq8\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:03 crc kubenswrapper[4935]: I1005 09:45:03.460005 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" event={"ID":"053eb908-340c-49d3-9ce3-0b126f4f3d44","Type":"ContainerDied","Data":"42edf8d3cee07146a2f2932083190709511868889085c1349305daa5945ef09e"} Oct 05 09:45:03 crc kubenswrapper[4935]: I1005 09:45:03.460105 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq" Oct 05 09:45:03 crc kubenswrapper[4935]: I1005 09:45:03.460706 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42edf8d3cee07146a2f2932083190709511868889085c1349305daa5945ef09e" Oct 05 09:45:03 crc kubenswrapper[4935]: I1005 09:45:03.912929 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz"] Oct 05 09:45:03 crc kubenswrapper[4935]: I1005 09:45:03.922190 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-r4hqz"] Oct 05 09:45:04 crc kubenswrapper[4935]: I1005 09:45:04.788067 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deae3a78-35c4-4fd6-bf16-b2cda1460dd7" path="/var/lib/kubelet/pods/deae3a78-35c4-4fd6-bf16-b2cda1460dd7/volumes" Oct 05 09:45:14 crc kubenswrapper[4935]: I1005 09:45:14.288867 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:45:14 crc kubenswrapper[4935]: I1005 09:45:14.289474 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:45:44 crc kubenswrapper[4935]: I1005 09:45:44.289576 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:45:44 crc kubenswrapper[4935]: I1005 09:45:44.290411 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:46:04 crc kubenswrapper[4935]: I1005 09:46:04.753373 4935 scope.go:117] "RemoveContainer" containerID="3d893bc44e401e3a801a1a454418d7deb79e1f7be7eebcd0d8b93cfb56bae961" Oct 05 09:46:14 crc kubenswrapper[4935]: I1005 09:46:14.289756 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:46:14 crc kubenswrapper[4935]: I1005 09:46:14.290476 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:46:14 crc kubenswrapper[4935]: I1005 09:46:14.290534 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:46:14 crc kubenswrapper[4935]: I1005 09:46:14.291690 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:46:14 crc kubenswrapper[4935]: I1005 09:46:14.291785 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" gracePeriod=600 Oct 05 09:46:14 crc kubenswrapper[4935]: E1005 09:46:14.440105 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:46:15 crc kubenswrapper[4935]: I1005 09:46:15.295379 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" exitCode=0 Oct 05 09:46:15 crc kubenswrapper[4935]: I1005 09:46:15.295452 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7"} Oct 05 09:46:15 crc kubenswrapper[4935]: I1005 09:46:15.295503 4935 scope.go:117] "RemoveContainer" containerID="c31b5e83aad13ab5aad5147c3c65b8e692b7949147dda7547b1fcc298444ac8b" Oct 05 09:46:15 crc kubenswrapper[4935]: I1005 09:46:15.296550 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:46:15 crc kubenswrapper[4935]: E1005 09:46:15.297039 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:46:28 crc kubenswrapper[4935]: I1005 09:46:28.778081 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:46:28 crc kubenswrapper[4935]: E1005 09:46:28.779442 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:46:41 crc kubenswrapper[4935]: I1005 09:46:41.777616 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:46:41 crc kubenswrapper[4935]: E1005 09:46:41.778557 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:46:53 crc kubenswrapper[4935]: I1005 09:46:53.778672 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:46:53 crc kubenswrapper[4935]: E1005 09:46:53.779951 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.418533 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dfnff"] Oct 05 09:46:58 crc kubenswrapper[4935]: E1005 09:46:58.419544 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053eb908-340c-49d3-9ce3-0b126f4f3d44" containerName="collect-profiles" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.419623 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="053eb908-340c-49d3-9ce3-0b126f4f3d44" containerName="collect-profiles" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.419842 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="053eb908-340c-49d3-9ce3-0b126f4f3d44" containerName="collect-profiles" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.421317 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.437225 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfnff"] Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.515953 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpdw9\" (UniqueName: \"kubernetes.io/projected/0650b5c6-093c-421e-bd64-1f63d4c4720c-kube-api-access-lpdw9\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.516064 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-utilities\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.516104 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-catalog-content\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.617376 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-utilities\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.617445 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-catalog-content\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.617594 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpdw9\" (UniqueName: \"kubernetes.io/projected/0650b5c6-093c-421e-bd64-1f63d4c4720c-kube-api-access-lpdw9\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.618080 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-catalog-content\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.618076 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-utilities\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.639983 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpdw9\" (UniqueName: \"kubernetes.io/projected/0650b5c6-093c-421e-bd64-1f63d4c4720c-kube-api-access-lpdw9\") pod \"redhat-operators-dfnff\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:58 crc kubenswrapper[4935]: I1005 09:46:58.749328 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:46:59 crc kubenswrapper[4935]: I1005 09:46:59.251682 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfnff"] Oct 05 09:46:59 crc kubenswrapper[4935]: I1005 09:46:59.860174 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnff" event={"ID":"0650b5c6-093c-421e-bd64-1f63d4c4720c","Type":"ContainerStarted","Data":"cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605"} Oct 05 09:46:59 crc kubenswrapper[4935]: I1005 09:46:59.860221 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnff" event={"ID":"0650b5c6-093c-421e-bd64-1f63d4c4720c","Type":"ContainerStarted","Data":"6e802e686e61579b76b8f13e63ce2a0833e183118c2446b6dd1a8ff4809f26c6"} Oct 05 09:47:00 crc kubenswrapper[4935]: I1005 09:47:00.874331 4935 generic.go:334] "Generic (PLEG): container finished" podID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerID="cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605" exitCode=0 Oct 05 09:47:00 crc kubenswrapper[4935]: I1005 09:47:00.874400 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnff" event={"ID":"0650b5c6-093c-421e-bd64-1f63d4c4720c","Type":"ContainerDied","Data":"cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605"} Oct 05 09:47:00 crc kubenswrapper[4935]: I1005 09:47:00.878059 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:47:00 crc kubenswrapper[4935]: I1005 09:47:00.878266 4935 generic.go:334] "Generic (PLEG): container finished" podID="8a8cc4c7-3c2c-4f19-9224-8d06432ca678" containerID="ecf203170c5eb69fbd47d67efa205ac6d0bf2c5f0988f21c49f726e85369f72b" exitCode=0 Oct 05 09:47:00 crc kubenswrapper[4935]: I1005 09:47:00.878344 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" event={"ID":"8a8cc4c7-3c2c-4f19-9224-8d06432ca678","Type":"ContainerDied","Data":"ecf203170c5eb69fbd47d67efa205ac6d0bf2c5f0988f21c49f726e85369f72b"} Oct 05 09:47:02 crc kubenswrapper[4935]: I1005 09:47:02.884588 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:47:02 crc kubenswrapper[4935]: I1005 09:47:02.901534 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnff" event={"ID":"0650b5c6-093c-421e-bd64-1f63d4c4720c","Type":"ContainerStarted","Data":"8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c"} Oct 05 09:47:02 crc kubenswrapper[4935]: I1005 09:47:02.903255 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" event={"ID":"8a8cc4c7-3c2c-4f19-9224-8d06432ca678","Type":"ContainerDied","Data":"89e5e9ca7b471c27a9a2a834a98b851cdf601dc4d7ac7ccd0f05f98d4d2c4767"} Oct 05 09:47:02 crc kubenswrapper[4935]: I1005 09:47:02.903289 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89e5e9ca7b471c27a9a2a834a98b851cdf601dc4d7ac7ccd0f05f98d4d2c4767" Oct 05 09:47:02 crc kubenswrapper[4935]: I1005 09:47:02.903333 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011509 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfkp8\" (UniqueName: \"kubernetes.io/projected/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-kube-api-access-pfkp8\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011600 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-0\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011631 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-inventory\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011655 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ceph\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011742 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-1\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011837 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-0\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011907 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-combined-ca-bundle\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011944 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ssh-key\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.011993 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-0\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.012101 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-1\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.012145 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-1\") pod \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\" (UID: \"8a8cc4c7-3c2c-4f19-9224-8d06432ca678\") " Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.026498 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ceph" (OuterVolumeSpecName: "ceph") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.026574 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-kube-api-access-pfkp8" (OuterVolumeSpecName: "kube-api-access-pfkp8") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "kube-api-access-pfkp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.030559 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.041183 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.043733 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.053093 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.057380 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.057967 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-inventory" (OuterVolumeSpecName: "inventory") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.058796 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114678 4935 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114709 4935 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ceph\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114720 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114731 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114739 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114749 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114757 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114765 4935 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.114773 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfkp8\" (UniqueName: \"kubernetes.io/projected/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-kube-api-access-pfkp8\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.160161 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.163771 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "8a8cc4c7-3c2c-4f19-9224-8d06432ca678" (UID: "8a8cc4c7-3c2c-4f19-9224-8d06432ca678"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.217170 4935 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:03 crc kubenswrapper[4935]: I1005 09:47:03.217203 4935 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a8cc4c7-3c2c-4f19-9224-8d06432ca678-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:04 crc kubenswrapper[4935]: I1005 09:47:04.931945 4935 generic.go:334] "Generic (PLEG): container finished" podID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerID="8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c" exitCode=0 Oct 05 09:47:04 crc kubenswrapper[4935]: I1005 09:47:04.932026 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnff" event={"ID":"0650b5c6-093c-421e-bd64-1f63d4c4720c","Type":"ContainerDied","Data":"8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c"} Oct 05 09:47:07 crc kubenswrapper[4935]: I1005 09:47:07.966466 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnff" event={"ID":"0650b5c6-093c-421e-bd64-1f63d4c4720c","Type":"ContainerStarted","Data":"196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3"} Oct 05 09:47:07 crc kubenswrapper[4935]: I1005 09:47:07.989977 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dfnff" podStartSLOduration=3.424727983 podStartE2EDuration="9.989947447s" podCreationTimestamp="2025-10-05 09:46:58 +0000 UTC" firstStartedPulling="2025-10-05 09:47:00.877430172 +0000 UTC m=+10454.760056682" lastFinishedPulling="2025-10-05 09:47:07.442649686 +0000 UTC m=+10461.325276146" observedRunningTime="2025-10-05 09:47:07.989487795 +0000 UTC m=+10461.872114265" watchObservedRunningTime="2025-10-05 09:47:07.989947447 +0000 UTC m=+10461.872573957" Oct 05 09:47:08 crc kubenswrapper[4935]: I1005 09:47:08.749814 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:47:08 crc kubenswrapper[4935]: I1005 09:47:08.750176 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:47:08 crc kubenswrapper[4935]: I1005 09:47:08.777544 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:47:08 crc kubenswrapper[4935]: E1005 09:47:08.777984 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:47:09 crc kubenswrapper[4935]: I1005 09:47:09.817108 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dfnff" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="registry-server" probeResult="failure" output=< Oct 05 09:47:09 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 09:47:09 crc kubenswrapper[4935]: > Oct 05 09:47:18 crc kubenswrapper[4935]: I1005 09:47:18.795417 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:47:18 crc kubenswrapper[4935]: I1005 09:47:18.840021 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:47:19 crc kubenswrapper[4935]: I1005 09:47:19.030323 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfnff"] Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.116281 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dfnff" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="registry-server" containerID="cri-o://196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3" gracePeriod=2 Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.719776 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.852011 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-catalog-content\") pod \"0650b5c6-093c-421e-bd64-1f63d4c4720c\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.852354 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpdw9\" (UniqueName: \"kubernetes.io/projected/0650b5c6-093c-421e-bd64-1f63d4c4720c-kube-api-access-lpdw9\") pod \"0650b5c6-093c-421e-bd64-1f63d4c4720c\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.852611 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-utilities\") pod \"0650b5c6-093c-421e-bd64-1f63d4c4720c\" (UID: \"0650b5c6-093c-421e-bd64-1f63d4c4720c\") " Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.853360 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-utilities" (OuterVolumeSpecName: "utilities") pod "0650b5c6-093c-421e-bd64-1f63d4c4720c" (UID: "0650b5c6-093c-421e-bd64-1f63d4c4720c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.859048 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0650b5c6-093c-421e-bd64-1f63d4c4720c-kube-api-access-lpdw9" (OuterVolumeSpecName: "kube-api-access-lpdw9") pod "0650b5c6-093c-421e-bd64-1f63d4c4720c" (UID: "0650b5c6-093c-421e-bd64-1f63d4c4720c"). InnerVolumeSpecName "kube-api-access-lpdw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.945083 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0650b5c6-093c-421e-bd64-1f63d4c4720c" (UID: "0650b5c6-093c-421e-bd64-1f63d4c4720c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.955538 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.955572 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0650b5c6-093c-421e-bd64-1f63d4c4720c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:20 crc kubenswrapper[4935]: I1005 09:47:20.955591 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpdw9\" (UniqueName: \"kubernetes.io/projected/0650b5c6-093c-421e-bd64-1f63d4c4720c-kube-api-access-lpdw9\") on node \"crc\" DevicePath \"\"" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.132015 4935 generic.go:334] "Generic (PLEG): container finished" podID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerID="196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3" exitCode=0 Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.132082 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnff" event={"ID":"0650b5c6-093c-421e-bd64-1f63d4c4720c","Type":"ContainerDied","Data":"196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3"} Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.132103 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfnff" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.132126 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnff" event={"ID":"0650b5c6-093c-421e-bd64-1f63d4c4720c","Type":"ContainerDied","Data":"6e802e686e61579b76b8f13e63ce2a0833e183118c2446b6dd1a8ff4809f26c6"} Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.132160 4935 scope.go:117] "RemoveContainer" containerID="196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.183916 4935 scope.go:117] "RemoveContainer" containerID="8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.189684 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfnff"] Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.199238 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dfnff"] Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.221413 4935 scope.go:117] "RemoveContainer" containerID="cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.273992 4935 scope.go:117] "RemoveContainer" containerID="196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3" Oct 05 09:47:21 crc kubenswrapper[4935]: E1005 09:47:21.274558 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3\": container with ID starting with 196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3 not found: ID does not exist" containerID="196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.274667 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3"} err="failed to get container status \"196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3\": rpc error: code = NotFound desc = could not find container \"196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3\": container with ID starting with 196756567f223cba4329a731240d973d3c2bbc4ffdefaa9278f3aebd1013ffa3 not found: ID does not exist" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.274715 4935 scope.go:117] "RemoveContainer" containerID="8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c" Oct 05 09:47:21 crc kubenswrapper[4935]: E1005 09:47:21.275193 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c\": container with ID starting with 8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c not found: ID does not exist" containerID="8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.275236 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c"} err="failed to get container status \"8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c\": rpc error: code = NotFound desc = could not find container \"8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c\": container with ID starting with 8821f7c7f175c00b832273cf27b5f61657ffd8a76cb93f5432e17dd7fd16649c not found: ID does not exist" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.275262 4935 scope.go:117] "RemoveContainer" containerID="cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605" Oct 05 09:47:21 crc kubenswrapper[4935]: E1005 09:47:21.275545 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605\": container with ID starting with cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605 not found: ID does not exist" containerID="cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605" Oct 05 09:47:21 crc kubenswrapper[4935]: I1005 09:47:21.275576 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605"} err="failed to get container status \"cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605\": rpc error: code = NotFound desc = could not find container \"cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605\": container with ID starting with cc25146f79f6d99198de0a9bd19e00e81091c382d5b33d40f08d554fc36c4605 not found: ID does not exist" Oct 05 09:47:22 crc kubenswrapper[4935]: I1005 09:47:22.777440 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:47:22 crc kubenswrapper[4935]: E1005 09:47:22.778220 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:47:22 crc kubenswrapper[4935]: I1005 09:47:22.792541 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" path="/var/lib/kubelet/pods/0650b5c6-093c-421e-bd64-1f63d4c4720c/volumes" Oct 05 09:47:35 crc kubenswrapper[4935]: I1005 09:47:35.777180 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:47:35 crc kubenswrapper[4935]: E1005 09:47:35.778063 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:47:48 crc kubenswrapper[4935]: I1005 09:47:48.778100 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:47:48 crc kubenswrapper[4935]: E1005 09:47:48.779216 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:47:48 crc kubenswrapper[4935]: E1005 09:47:48.943088 4935 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.59:57218->38.102.83.59:37063: read tcp 38.102.83.59:57218->38.102.83.59:37063: read: connection reset by peer Oct 05 09:47:48 crc kubenswrapper[4935]: E1005 09:47:48.943146 4935 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.59:57218->38.102.83.59:37063: write tcp 38.102.83.59:57218->38.102.83.59:37063: write: broken pipe Oct 05 09:48:03 crc kubenswrapper[4935]: I1005 09:48:03.777946 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:48:03 crc kubenswrapper[4935]: E1005 09:48:03.779173 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:48:18 crc kubenswrapper[4935]: I1005 09:48:18.777646 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:48:18 crc kubenswrapper[4935]: E1005 09:48:18.778471 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:48:29 crc kubenswrapper[4935]: I1005 09:48:29.777809 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:48:29 crc kubenswrapper[4935]: E1005 09:48:29.778606 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:48:29 crc kubenswrapper[4935]: I1005 09:48:29.992550 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 09:48:29 crc kubenswrapper[4935]: I1005 09:48:29.992979 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="ebe6c9a7-c4a8-454a-973d-b6a1d849fdda" containerName="adoption" containerID="cri-o://9818a1fd3c4ffb64ca1593726de0175068edda0a60422d2666f1be36fc7a773a" gracePeriod=30 Oct 05 09:48:40 crc kubenswrapper[4935]: I1005 09:48:40.777973 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:48:40 crc kubenswrapper[4935]: E1005 09:48:40.778738 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:48:51 crc kubenswrapper[4935]: I1005 09:48:51.777881 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:48:51 crc kubenswrapper[4935]: E1005 09:48:51.778611 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.382811 4935 generic.go:334] "Generic (PLEG): container finished" podID="ebe6c9a7-c4a8-454a-973d-b6a1d849fdda" containerID="9818a1fd3c4ffb64ca1593726de0175068edda0a60422d2666f1be36fc7a773a" exitCode=137 Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.383012 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda","Type":"ContainerDied","Data":"9818a1fd3c4ffb64ca1593726de0175068edda0a60422d2666f1be36fc7a773a"} Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.642168 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.797975 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46f2r\" (UniqueName: \"kubernetes.io/projected/ebe6c9a7-c4a8-454a-973d-b6a1d849fdda-kube-api-access-46f2r\") pod \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") " Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.800004 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\") pod \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\" (UID: \"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda\") " Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.809586 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebe6c9a7-c4a8-454a-973d-b6a1d849fdda-kube-api-access-46f2r" (OuterVolumeSpecName: "kube-api-access-46f2r") pod "ebe6c9a7-c4a8-454a-973d-b6a1d849fdda" (UID: "ebe6c9a7-c4a8-454a-973d-b6a1d849fdda"). InnerVolumeSpecName "kube-api-access-46f2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.828704 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5" (OuterVolumeSpecName: "mariadb-data") pod "ebe6c9a7-c4a8-454a-973d-b6a1d849fdda" (UID: "ebe6c9a7-c4a8-454a-973d-b6a1d849fdda"). InnerVolumeSpecName "pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.906211 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46f2r\" (UniqueName: \"kubernetes.io/projected/ebe6c9a7-c4a8-454a-973d-b6a1d849fdda-kube-api-access-46f2r\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.906280 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\") on node \"crc\" " Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.934829 4935 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 09:49:00 crc kubenswrapper[4935]: I1005 09:49:00.935111 4935 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5") on node "crc" Oct 05 09:49:01 crc kubenswrapper[4935]: I1005 09:49:01.008024 4935 reconciler_common.go:293] "Volume detached for volume \"pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2d1a638-e0b0-49cc-9380-22c78b3a38f5\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:01 crc kubenswrapper[4935]: I1005 09:49:01.396165 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"ebe6c9a7-c4a8-454a-973d-b6a1d849fdda","Type":"ContainerDied","Data":"7193eaff12ec2721972f59bddcc66171bca0b7c2f10275846f7579c66aeddc57"} Oct 05 09:49:01 crc kubenswrapper[4935]: I1005 09:49:01.396195 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 05 09:49:01 crc kubenswrapper[4935]: I1005 09:49:01.396241 4935 scope.go:117] "RemoveContainer" containerID="9818a1fd3c4ffb64ca1593726de0175068edda0a60422d2666f1be36fc7a773a" Oct 05 09:49:01 crc kubenswrapper[4935]: I1005 09:49:01.440202 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 09:49:01 crc kubenswrapper[4935]: I1005 09:49:01.449386 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 09:49:02 crc kubenswrapper[4935]: I1005 09:49:02.184789 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 09:49:02 crc kubenswrapper[4935]: I1005 09:49:02.185535 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="ec474e79-9762-42c6-a931-bfa5896f7f00" containerName="adoption" containerID="cri-o://17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66" gracePeriod=30 Oct 05 09:49:02 crc kubenswrapper[4935]: I1005 09:49:02.793330 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebe6c9a7-c4a8-454a-973d-b6a1d849fdda" path="/var/lib/kubelet/pods/ebe6c9a7-c4a8-454a-973d-b6a1d849fdda/volumes" Oct 05 09:49:05 crc kubenswrapper[4935]: I1005 09:49:05.777051 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:49:05 crc kubenswrapper[4935]: E1005 09:49:05.777804 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:49:18 crc kubenswrapper[4935]: I1005 09:49:18.794739 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:49:18 crc kubenswrapper[4935]: E1005 09:49:18.809394 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.796245 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-25ljk"] Oct 05 09:49:24 crc kubenswrapper[4935]: E1005 09:49:24.797618 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a8cc4c7-3c2c-4f19-9224-8d06432ca678" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.797648 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a8cc4c7-3c2c-4f19-9224-8d06432ca678" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:49:24 crc kubenswrapper[4935]: E1005 09:49:24.797677 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="extract-content" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.797692 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="extract-content" Oct 05 09:49:24 crc kubenswrapper[4935]: E1005 09:49:24.797738 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="registry-server" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.797751 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="registry-server" Oct 05 09:49:24 crc kubenswrapper[4935]: E1005 09:49:24.797797 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="extract-utilities" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.797809 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="extract-utilities" Oct 05 09:49:24 crc kubenswrapper[4935]: E1005 09:49:24.797827 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebe6c9a7-c4a8-454a-973d-b6a1d849fdda" containerName="adoption" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.797840 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebe6c9a7-c4a8-454a-973d-b6a1d849fdda" containerName="adoption" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.798262 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebe6c9a7-c4a8-454a-973d-b6a1d849fdda" containerName="adoption" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.798319 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a8cc4c7-3c2c-4f19-9224-8d06432ca678" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.798354 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0650b5c6-093c-421e-bd64-1f63d4c4720c" containerName="registry-server" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.801836 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.809432 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-25ljk"] Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.897512 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-catalog-content\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.897667 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvv7t\" (UniqueName: \"kubernetes.io/projected/bb38681a-675b-4dc3-bb92-083d6dbf8231-kube-api-access-qvv7t\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.897851 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-utilities\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.999783 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvv7t\" (UniqueName: \"kubernetes.io/projected/bb38681a-675b-4dc3-bb92-083d6dbf8231-kube-api-access-qvv7t\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:24 crc kubenswrapper[4935]: I1005 09:49:24.999830 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-utilities\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:25 crc kubenswrapper[4935]: I1005 09:49:25.000492 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-utilities\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:25 crc kubenswrapper[4935]: I1005 09:49:25.004077 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-catalog-content\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:25 crc kubenswrapper[4935]: I1005 09:49:25.004518 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-catalog-content\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:25 crc kubenswrapper[4935]: I1005 09:49:25.028534 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvv7t\" (UniqueName: \"kubernetes.io/projected/bb38681a-675b-4dc3-bb92-083d6dbf8231-kube-api-access-qvv7t\") pod \"certified-operators-25ljk\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:25 crc kubenswrapper[4935]: I1005 09:49:25.120987 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:25 crc kubenswrapper[4935]: W1005 09:49:25.677613 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb38681a_675b_4dc3_bb92_083d6dbf8231.slice/crio-25699184bcb7e53c9264ff9ab0f46261433234c3ddcec0cb4f3cf2ca2b16309f WatchSource:0}: Error finding container 25699184bcb7e53c9264ff9ab0f46261433234c3ddcec0cb4f3cf2ca2b16309f: Status 404 returned error can't find the container with id 25699184bcb7e53c9264ff9ab0f46261433234c3ddcec0cb4f3cf2ca2b16309f Oct 05 09:49:25 crc kubenswrapper[4935]: I1005 09:49:25.685479 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-25ljk"] Oct 05 09:49:25 crc kubenswrapper[4935]: I1005 09:49:25.698272 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25ljk" event={"ID":"bb38681a-675b-4dc3-bb92-083d6dbf8231","Type":"ContainerStarted","Data":"25699184bcb7e53c9264ff9ab0f46261433234c3ddcec0cb4f3cf2ca2b16309f"} Oct 05 09:49:26 crc kubenswrapper[4935]: I1005 09:49:26.714740 4935 generic.go:334] "Generic (PLEG): container finished" podID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerID="57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7" exitCode=0 Oct 05 09:49:26 crc kubenswrapper[4935]: I1005 09:49:26.715644 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25ljk" event={"ID":"bb38681a-675b-4dc3-bb92-083d6dbf8231","Type":"ContainerDied","Data":"57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7"} Oct 05 09:49:28 crc kubenswrapper[4935]: I1005 09:49:28.743058 4935 generic.go:334] "Generic (PLEG): container finished" podID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerID="d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb" exitCode=0 Oct 05 09:49:28 crc kubenswrapper[4935]: I1005 09:49:28.743165 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25ljk" event={"ID":"bb38681a-675b-4dc3-bb92-083d6dbf8231","Type":"ContainerDied","Data":"d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb"} Oct 05 09:49:29 crc kubenswrapper[4935]: I1005 09:49:29.756630 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25ljk" event={"ID":"bb38681a-675b-4dc3-bb92-083d6dbf8231","Type":"ContainerStarted","Data":"ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37"} Oct 05 09:49:29 crc kubenswrapper[4935]: I1005 09:49:29.788772 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-25ljk" podStartSLOduration=3.2781818830000002 podStartE2EDuration="5.788756142s" podCreationTimestamp="2025-10-05 09:49:24 +0000 UTC" firstStartedPulling="2025-10-05 09:49:26.719993413 +0000 UTC m=+10600.602619903" lastFinishedPulling="2025-10-05 09:49:29.230567662 +0000 UTC m=+10603.113194162" observedRunningTime="2025-10-05 09:49:29.782577858 +0000 UTC m=+10603.665204318" watchObservedRunningTime="2025-10-05 09:49:29.788756142 +0000 UTC m=+10603.671382602" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.797527 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.797590 4935 generic.go:334] "Generic (PLEG): container finished" podID="ec474e79-9762-42c6-a931-bfa5896f7f00" containerID="17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66" exitCode=137 Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.797623 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"ec474e79-9762-42c6-a931-bfa5896f7f00","Type":"ContainerDied","Data":"17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66"} Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.798190 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"ec474e79-9762-42c6-a931-bfa5896f7f00","Type":"ContainerDied","Data":"27f47c837e5a78c1d78398006072c5f31b7f66c33fe7c1605fc6cb22f70a2e88"} Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.798225 4935 scope.go:117] "RemoveContainer" containerID="17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.834188 4935 scope.go:117] "RemoveContainer" containerID="17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66" Oct 05 09:49:32 crc kubenswrapper[4935]: E1005 09:49:32.834835 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66\": container with ID starting with 17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66 not found: ID does not exist" containerID="17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.834878 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66"} err="failed to get container status \"17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66\": rpc error: code = NotFound desc = could not find container \"17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66\": container with ID starting with 17fd4138fd8bb7418a88146039b5ef48fcb71bcad750013c94c90e9257220d66 not found: ID does not exist" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.887361 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/ec474e79-9762-42c6-a931-bfa5896f7f00-ovn-data-cert\") pod \"ec474e79-9762-42c6-a931-bfa5896f7f00\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.888255 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\") pod \"ec474e79-9762-42c6-a931-bfa5896f7f00\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.888389 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hx96\" (UniqueName: \"kubernetes.io/projected/ec474e79-9762-42c6-a931-bfa5896f7f00-kube-api-access-9hx96\") pod \"ec474e79-9762-42c6-a931-bfa5896f7f00\" (UID: \"ec474e79-9762-42c6-a931-bfa5896f7f00\") " Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.893799 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec474e79-9762-42c6-a931-bfa5896f7f00-kube-api-access-9hx96" (OuterVolumeSpecName: "kube-api-access-9hx96") pod "ec474e79-9762-42c6-a931-bfa5896f7f00" (UID: "ec474e79-9762-42c6-a931-bfa5896f7f00"). InnerVolumeSpecName "kube-api-access-9hx96". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.893870 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec474e79-9762-42c6-a931-bfa5896f7f00-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "ec474e79-9762-42c6-a931-bfa5896f7f00" (UID: "ec474e79-9762-42c6-a931-bfa5896f7f00"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.902517 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b" (OuterVolumeSpecName: "ovn-data") pod "ec474e79-9762-42c6-a931-bfa5896f7f00" (UID: "ec474e79-9762-42c6-a931-bfa5896f7f00"). InnerVolumeSpecName "pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.990298 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\") on node \"crc\" " Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.990334 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hx96\" (UniqueName: \"kubernetes.io/projected/ec474e79-9762-42c6-a931-bfa5896f7f00-kube-api-access-9hx96\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:32 crc kubenswrapper[4935]: I1005 09:49:32.990345 4935 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/ec474e79-9762-42c6-a931-bfa5896f7f00-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:33 crc kubenswrapper[4935]: I1005 09:49:33.013876 4935 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 09:49:33 crc kubenswrapper[4935]: I1005 09:49:33.014052 4935 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b") on node "crc" Oct 05 09:49:33 crc kubenswrapper[4935]: I1005 09:49:33.092620 4935 reconciler_common.go:293] "Volume detached for volume \"pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-137df2cc-da5f-4023-8329-e4d2c57dc41b\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:33 crc kubenswrapper[4935]: I1005 09:49:33.779006 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:49:33 crc kubenswrapper[4935]: E1005 09:49:33.779431 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:49:33 crc kubenswrapper[4935]: I1005 09:49:33.815346 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 05 09:49:33 crc kubenswrapper[4935]: I1005 09:49:33.879866 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 09:49:33 crc kubenswrapper[4935]: I1005 09:49:33.898999 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 09:49:34 crc kubenswrapper[4935]: I1005 09:49:34.796800 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec474e79-9762-42c6-a931-bfa5896f7f00" path="/var/lib/kubelet/pods/ec474e79-9762-42c6-a931-bfa5896f7f00/volumes" Oct 05 09:49:35 crc kubenswrapper[4935]: I1005 09:49:35.122016 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:35 crc kubenswrapper[4935]: I1005 09:49:35.122086 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:35 crc kubenswrapper[4935]: I1005 09:49:35.202421 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:35 crc kubenswrapper[4935]: I1005 09:49:35.892570 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:35 crc kubenswrapper[4935]: I1005 09:49:35.954597 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-25ljk"] Oct 05 09:49:37 crc kubenswrapper[4935]: I1005 09:49:37.862831 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-25ljk" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerName="registry-server" containerID="cri-o://ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37" gracePeriod=2 Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.436003 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.530125 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-utilities\") pod \"bb38681a-675b-4dc3-bb92-083d6dbf8231\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.530421 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-catalog-content\") pod \"bb38681a-675b-4dc3-bb92-083d6dbf8231\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.530550 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvv7t\" (UniqueName: \"kubernetes.io/projected/bb38681a-675b-4dc3-bb92-083d6dbf8231-kube-api-access-qvv7t\") pod \"bb38681a-675b-4dc3-bb92-083d6dbf8231\" (UID: \"bb38681a-675b-4dc3-bb92-083d6dbf8231\") " Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.531160 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-utilities" (OuterVolumeSpecName: "utilities") pod "bb38681a-675b-4dc3-bb92-083d6dbf8231" (UID: "bb38681a-675b-4dc3-bb92-083d6dbf8231"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.538776 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb38681a-675b-4dc3-bb92-083d6dbf8231-kube-api-access-qvv7t" (OuterVolumeSpecName: "kube-api-access-qvv7t") pod "bb38681a-675b-4dc3-bb92-083d6dbf8231" (UID: "bb38681a-675b-4dc3-bb92-083d6dbf8231"). InnerVolumeSpecName "kube-api-access-qvv7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.581670 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb38681a-675b-4dc3-bb92-083d6dbf8231" (UID: "bb38681a-675b-4dc3-bb92-083d6dbf8231"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.633131 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.633169 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvv7t\" (UniqueName: \"kubernetes.io/projected/bb38681a-675b-4dc3-bb92-083d6dbf8231-kube-api-access-qvv7t\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.633181 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb38681a-675b-4dc3-bb92-083d6dbf8231-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.876721 4935 generic.go:334] "Generic (PLEG): container finished" podID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerID="ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37" exitCode=0 Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.876767 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25ljk" event={"ID":"bb38681a-675b-4dc3-bb92-083d6dbf8231","Type":"ContainerDied","Data":"ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37"} Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.876794 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25ljk" event={"ID":"bb38681a-675b-4dc3-bb92-083d6dbf8231","Type":"ContainerDied","Data":"25699184bcb7e53c9264ff9ab0f46261433234c3ddcec0cb4f3cf2ca2b16309f"} Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.876811 4935 scope.go:117] "RemoveContainer" containerID="ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.876840 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-25ljk" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.912273 4935 scope.go:117] "RemoveContainer" containerID="d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb" Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.919841 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-25ljk"] Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.937880 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-25ljk"] Oct 05 09:49:38 crc kubenswrapper[4935]: I1005 09:49:38.954042 4935 scope.go:117] "RemoveContainer" containerID="57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7" Oct 05 09:49:39 crc kubenswrapper[4935]: I1005 09:49:39.003177 4935 scope.go:117] "RemoveContainer" containerID="ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37" Oct 05 09:49:39 crc kubenswrapper[4935]: E1005 09:49:39.003623 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37\": container with ID starting with ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37 not found: ID does not exist" containerID="ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37" Oct 05 09:49:39 crc kubenswrapper[4935]: I1005 09:49:39.003672 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37"} err="failed to get container status \"ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37\": rpc error: code = NotFound desc = could not find container \"ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37\": container with ID starting with ed41dafe13a5b82dac440f6da547a1693befb8b56dee9866414b2508603bbd37 not found: ID does not exist" Oct 05 09:49:39 crc kubenswrapper[4935]: I1005 09:49:39.003701 4935 scope.go:117] "RemoveContainer" containerID="d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb" Oct 05 09:49:39 crc kubenswrapper[4935]: E1005 09:49:39.004052 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb\": container with ID starting with d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb not found: ID does not exist" containerID="d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb" Oct 05 09:49:39 crc kubenswrapper[4935]: I1005 09:49:39.004081 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb"} err="failed to get container status \"d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb\": rpc error: code = NotFound desc = could not find container \"d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb\": container with ID starting with d0f35da33641706f0a9dce95373f0fde976d8e5f1cfc17d29d625fae846990bb not found: ID does not exist" Oct 05 09:49:39 crc kubenswrapper[4935]: I1005 09:49:39.004093 4935 scope.go:117] "RemoveContainer" containerID="57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7" Oct 05 09:49:39 crc kubenswrapper[4935]: E1005 09:49:39.004299 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7\": container with ID starting with 57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7 not found: ID does not exist" containerID="57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7" Oct 05 09:49:39 crc kubenswrapper[4935]: I1005 09:49:39.004319 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7"} err="failed to get container status \"57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7\": rpc error: code = NotFound desc = could not find container \"57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7\": container with ID starting with 57b8e06d043dcc3ccfde4d2fbe95b2e9afab154a990dc325cc35375e8ceecdd7 not found: ID does not exist" Oct 05 09:49:40 crc kubenswrapper[4935]: I1005 09:49:40.791986 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" path="/var/lib/kubelet/pods/bb38681a-675b-4dc3-bb92-083d6dbf8231/volumes" Oct 05 09:49:47 crc kubenswrapper[4935]: I1005 09:49:47.777877 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:49:47 crc kubenswrapper[4935]: E1005 09:49:47.778538 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:50:01 crc kubenswrapper[4935]: I1005 09:50:01.776858 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:50:01 crc kubenswrapper[4935]: E1005 09:50:01.777751 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:50:14 crc kubenswrapper[4935]: I1005 09:50:14.777403 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:50:14 crc kubenswrapper[4935]: E1005 09:50:14.778273 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:50:26 crc kubenswrapper[4935]: I1005 09:50:26.806458 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:50:26 crc kubenswrapper[4935]: E1005 09:50:26.807539 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:50:40 crc kubenswrapper[4935]: I1005 09:50:40.780132 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:50:40 crc kubenswrapper[4935]: E1005 09:50:40.783774 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.008688 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8twd8"] Oct 05 09:50:41 crc kubenswrapper[4935]: E1005 09:50:41.009417 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerName="extract-content" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.009449 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerName="extract-content" Oct 05 09:50:41 crc kubenswrapper[4935]: E1005 09:50:41.009465 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerName="registry-server" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.009478 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerName="registry-server" Oct 05 09:50:41 crc kubenswrapper[4935]: E1005 09:50:41.009512 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerName="extract-utilities" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.009528 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerName="extract-utilities" Oct 05 09:50:41 crc kubenswrapper[4935]: E1005 09:50:41.009548 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec474e79-9762-42c6-a931-bfa5896f7f00" containerName="adoption" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.009561 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec474e79-9762-42c6-a931-bfa5896f7f00" containerName="adoption" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.009995 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec474e79-9762-42c6-a931-bfa5896f7f00" containerName="adoption" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.010056 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb38681a-675b-4dc3-bb92-083d6dbf8231" containerName="registry-server" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.012793 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.033481 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twd8"] Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.142685 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fr7g\" (UniqueName: \"kubernetes.io/projected/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-kube-api-access-6fr7g\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.142749 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-catalog-content\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.143394 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-utilities\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.245221 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-utilities\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.245314 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fr7g\" (UniqueName: \"kubernetes.io/projected/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-kube-api-access-6fr7g\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.245334 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-catalog-content\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.245780 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-catalog-content\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.246908 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-utilities\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.275250 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fr7g\" (UniqueName: \"kubernetes.io/projected/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-kube-api-access-6fr7g\") pod \"redhat-marketplace-8twd8\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.349979 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:41 crc kubenswrapper[4935]: I1005 09:50:41.818533 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twd8"] Oct 05 09:50:41 crc kubenswrapper[4935]: W1005 09:50:41.828119 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf505c30a_8e93_4bf4_96f8_d9d649f3e7ae.slice/crio-467c35d5796ed164346b0ac673ec874ab789e24bba32660055c57744cb1570d2 WatchSource:0}: Error finding container 467c35d5796ed164346b0ac673ec874ab789e24bba32660055c57744cb1570d2: Status 404 returned error can't find the container with id 467c35d5796ed164346b0ac673ec874ab789e24bba32660055c57744cb1570d2 Oct 05 09:50:42 crc kubenswrapper[4935]: I1005 09:50:42.710525 4935 generic.go:334] "Generic (PLEG): container finished" podID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerID="4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd" exitCode=0 Oct 05 09:50:42 crc kubenswrapper[4935]: I1005 09:50:42.710576 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twd8" event={"ID":"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae","Type":"ContainerDied","Data":"4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd"} Oct 05 09:50:42 crc kubenswrapper[4935]: I1005 09:50:42.710608 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twd8" event={"ID":"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae","Type":"ContainerStarted","Data":"467c35d5796ed164346b0ac673ec874ab789e24bba32660055c57744cb1570d2"} Oct 05 09:50:43 crc kubenswrapper[4935]: I1005 09:50:43.730623 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twd8" event={"ID":"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae","Type":"ContainerStarted","Data":"82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a"} Oct 05 09:50:44 crc kubenswrapper[4935]: I1005 09:50:44.745811 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twd8" event={"ID":"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae","Type":"ContainerDied","Data":"82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a"} Oct 05 09:50:44 crc kubenswrapper[4935]: I1005 09:50:44.745620 4935 generic.go:334] "Generic (PLEG): container finished" podID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerID="82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a" exitCode=0 Oct 05 09:50:45 crc kubenswrapper[4935]: I1005 09:50:45.763708 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twd8" event={"ID":"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae","Type":"ContainerStarted","Data":"2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612"} Oct 05 09:50:45 crc kubenswrapper[4935]: I1005 09:50:45.788458 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8twd8" podStartSLOduration=3.329452378 podStartE2EDuration="5.788439321s" podCreationTimestamp="2025-10-05 09:50:40 +0000 UTC" firstStartedPulling="2025-10-05 09:50:42.714142726 +0000 UTC m=+10676.596769226" lastFinishedPulling="2025-10-05 09:50:45.173129709 +0000 UTC m=+10679.055756169" observedRunningTime="2025-10-05 09:50:45.786763366 +0000 UTC m=+10679.669389846" watchObservedRunningTime="2025-10-05 09:50:45.788439321 +0000 UTC m=+10679.671065791" Oct 05 09:50:51 crc kubenswrapper[4935]: I1005 09:50:51.350991 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:51 crc kubenswrapper[4935]: I1005 09:50:51.351996 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:51 crc kubenswrapper[4935]: I1005 09:50:51.430358 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:51 crc kubenswrapper[4935]: I1005 09:50:51.898196 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:51 crc kubenswrapper[4935]: I1005 09:50:51.943832 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twd8"] Oct 05 09:50:53 crc kubenswrapper[4935]: I1005 09:50:53.866136 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8twd8" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerName="registry-server" containerID="cri-o://2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612" gracePeriod=2 Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.479663 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.554443 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-catalog-content\") pod \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.554518 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-utilities\") pod \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.554720 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fr7g\" (UniqueName: \"kubernetes.io/projected/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-kube-api-access-6fr7g\") pod \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\" (UID: \"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae\") " Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.555495 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-utilities" (OuterVolumeSpecName: "utilities") pod "f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" (UID: "f505c30a-8e93-4bf4-96f8-d9d649f3e7ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.560090 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-kube-api-access-6fr7g" (OuterVolumeSpecName: "kube-api-access-6fr7g") pod "f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" (UID: "f505c30a-8e93-4bf4-96f8-d9d649f3e7ae"). InnerVolumeSpecName "kube-api-access-6fr7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.567860 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" (UID: "f505c30a-8e93-4bf4-96f8-d9d649f3e7ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.657400 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.657684 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.657772 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fr7g\" (UniqueName: \"kubernetes.io/projected/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae-kube-api-access-6fr7g\") on node \"crc\" DevicePath \"\"" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.884814 4935 generic.go:334] "Generic (PLEG): container finished" podID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerID="2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612" exitCode=0 Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.884959 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8twd8" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.884911 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twd8" event={"ID":"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae","Type":"ContainerDied","Data":"2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612"} Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.885114 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twd8" event={"ID":"f505c30a-8e93-4bf4-96f8-d9d649f3e7ae","Type":"ContainerDied","Data":"467c35d5796ed164346b0ac673ec874ab789e24bba32660055c57744cb1570d2"} Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.885138 4935 scope.go:117] "RemoveContainer" containerID="2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.932935 4935 scope.go:117] "RemoveContainer" containerID="82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a" Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.933635 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twd8"] Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.955956 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twd8"] Oct 05 09:50:54 crc kubenswrapper[4935]: I1005 09:50:54.973417 4935 scope.go:117] "RemoveContainer" containerID="4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd" Oct 05 09:50:55 crc kubenswrapper[4935]: I1005 09:50:55.024218 4935 scope.go:117] "RemoveContainer" containerID="2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612" Oct 05 09:50:55 crc kubenswrapper[4935]: E1005 09:50:55.024864 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612\": container with ID starting with 2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612 not found: ID does not exist" containerID="2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612" Oct 05 09:50:55 crc kubenswrapper[4935]: I1005 09:50:55.024973 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612"} err="failed to get container status \"2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612\": rpc error: code = NotFound desc = could not find container \"2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612\": container with ID starting with 2c4a5fe8726662eddfe3ccb5b4c1f6ff10efbbeffc2e79a9556919b8726ff612 not found: ID does not exist" Oct 05 09:50:55 crc kubenswrapper[4935]: I1005 09:50:55.025052 4935 scope.go:117] "RemoveContainer" containerID="82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a" Oct 05 09:50:55 crc kubenswrapper[4935]: E1005 09:50:55.026207 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a\": container with ID starting with 82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a not found: ID does not exist" containerID="82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a" Oct 05 09:50:55 crc kubenswrapper[4935]: I1005 09:50:55.026275 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a"} err="failed to get container status \"82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a\": rpc error: code = NotFound desc = could not find container \"82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a\": container with ID starting with 82ab2f91c93597b1e98e1edcdfcf66da3732ad40837adcac51bab750cce1069a not found: ID does not exist" Oct 05 09:50:55 crc kubenswrapper[4935]: I1005 09:50:55.026320 4935 scope.go:117] "RemoveContainer" containerID="4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd" Oct 05 09:50:55 crc kubenswrapper[4935]: E1005 09:50:55.027605 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd\": container with ID starting with 4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd not found: ID does not exist" containerID="4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd" Oct 05 09:50:55 crc kubenswrapper[4935]: I1005 09:50:55.027677 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd"} err="failed to get container status \"4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd\": rpc error: code = NotFound desc = could not find container \"4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd\": container with ID starting with 4a4135980a5ba34aa9884a84896b6a6ad14e007d60cdd64b671e4a29df3d32bd not found: ID does not exist" Oct 05 09:50:55 crc kubenswrapper[4935]: I1005 09:50:55.777651 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:50:55 crc kubenswrapper[4935]: E1005 09:50:55.778507 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:50:56 crc kubenswrapper[4935]: I1005 09:50:56.789470 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" path="/var/lib/kubelet/pods/f505c30a-8e93-4bf4-96f8-d9d649f3e7ae/volumes" Oct 05 09:51:07 crc kubenswrapper[4935]: I1005 09:51:07.777185 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:51:07 crc kubenswrapper[4935]: E1005 09:51:07.778041 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:51:21 crc kubenswrapper[4935]: I1005 09:51:21.778138 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:51:22 crc kubenswrapper[4935]: I1005 09:51:22.245476 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"58c70cc3730353e78569d63ee4affa370bf03d9815dc648f7d1464254f04339b"} Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.308478 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bxp77"] Oct 05 09:51:48 crc kubenswrapper[4935]: E1005 09:51:48.310066 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerName="extract-utilities" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.310235 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerName="extract-utilities" Oct 05 09:51:48 crc kubenswrapper[4935]: E1005 09:51:48.310320 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerName="registry-server" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.310338 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerName="registry-server" Oct 05 09:51:48 crc kubenswrapper[4935]: E1005 09:51:48.310404 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerName="extract-content" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.310423 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerName="extract-content" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.310982 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f505c30a-8e93-4bf4-96f8-d9d649f3e7ae" containerName="registry-server" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.314133 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.326964 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bxp77"] Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.386815 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtl5b\" (UniqueName: \"kubernetes.io/projected/3270814b-f542-4706-bfaa-63403d036152-kube-api-access-dtl5b\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.387557 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-utilities\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.387954 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-catalog-content\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.490158 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-catalog-content\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.490966 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtl5b\" (UniqueName: \"kubernetes.io/projected/3270814b-f542-4706-bfaa-63403d036152-kube-api-access-dtl5b\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.491037 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-utilities\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.491098 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-catalog-content\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.491588 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-utilities\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.510934 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtl5b\" (UniqueName: \"kubernetes.io/projected/3270814b-f542-4706-bfaa-63403d036152-kube-api-access-dtl5b\") pod \"community-operators-bxp77\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:48 crc kubenswrapper[4935]: I1005 09:51:48.657493 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:49 crc kubenswrapper[4935]: I1005 09:51:49.220369 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bxp77"] Oct 05 09:51:49 crc kubenswrapper[4935]: I1005 09:51:49.649870 4935 generic.go:334] "Generic (PLEG): container finished" podID="3270814b-f542-4706-bfaa-63403d036152" containerID="7c1d3754e0068018253e448791e276bb728cc5f4ec234e74500c5a1422707f21" exitCode=0 Oct 05 09:51:49 crc kubenswrapper[4935]: I1005 09:51:49.649974 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bxp77" event={"ID":"3270814b-f542-4706-bfaa-63403d036152","Type":"ContainerDied","Data":"7c1d3754e0068018253e448791e276bb728cc5f4ec234e74500c5a1422707f21"} Oct 05 09:51:49 crc kubenswrapper[4935]: I1005 09:51:49.650318 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bxp77" event={"ID":"3270814b-f542-4706-bfaa-63403d036152","Type":"ContainerStarted","Data":"0bd6300eec2cd6261d179014eabf45e91385902b7a11cc4ec0125104c0404992"} Oct 05 09:51:50 crc kubenswrapper[4935]: I1005 09:51:50.666596 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bxp77" event={"ID":"3270814b-f542-4706-bfaa-63403d036152","Type":"ContainerStarted","Data":"6997c33ea25080a5167f599c202eb35fd9e70440b9bfd51479ef9d8c77304f8d"} Oct 05 09:51:51 crc kubenswrapper[4935]: I1005 09:51:51.688226 4935 generic.go:334] "Generic (PLEG): container finished" podID="3270814b-f542-4706-bfaa-63403d036152" containerID="6997c33ea25080a5167f599c202eb35fd9e70440b9bfd51479ef9d8c77304f8d" exitCode=0 Oct 05 09:51:51 crc kubenswrapper[4935]: I1005 09:51:51.688609 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bxp77" event={"ID":"3270814b-f542-4706-bfaa-63403d036152","Type":"ContainerDied","Data":"6997c33ea25080a5167f599c202eb35fd9e70440b9bfd51479ef9d8c77304f8d"} Oct 05 09:51:52 crc kubenswrapper[4935]: I1005 09:51:52.715614 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bxp77" event={"ID":"3270814b-f542-4706-bfaa-63403d036152","Type":"ContainerStarted","Data":"11d0f23969461ca00d3e586d504d52406021f5ad6cc06b657c3641de7da30974"} Oct 05 09:51:52 crc kubenswrapper[4935]: I1005 09:51:52.737155 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bxp77" podStartSLOduration=2.080236474 podStartE2EDuration="4.737139974s" podCreationTimestamp="2025-10-05 09:51:48 +0000 UTC" firstStartedPulling="2025-10-05 09:51:49.651758216 +0000 UTC m=+10743.534384676" lastFinishedPulling="2025-10-05 09:51:52.308661706 +0000 UTC m=+10746.191288176" observedRunningTime="2025-10-05 09:51:52.735669545 +0000 UTC m=+10746.618296035" watchObservedRunningTime="2025-10-05 09:51:52.737139974 +0000 UTC m=+10746.619766434" Oct 05 09:51:58 crc kubenswrapper[4935]: I1005 09:51:58.658500 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:58 crc kubenswrapper[4935]: I1005 09:51:58.660116 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:58 crc kubenswrapper[4935]: I1005 09:51:58.714193 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:51:58 crc kubenswrapper[4935]: I1005 09:51:58.888041 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:52:01 crc kubenswrapper[4935]: I1005 09:52:01.676512 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bxp77"] Oct 05 09:52:01 crc kubenswrapper[4935]: I1005 09:52:01.677079 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bxp77" podUID="3270814b-f542-4706-bfaa-63403d036152" containerName="registry-server" containerID="cri-o://11d0f23969461ca00d3e586d504d52406021f5ad6cc06b657c3641de7da30974" gracePeriod=2 Oct 05 09:52:01 crc kubenswrapper[4935]: I1005 09:52:01.853649 4935 generic.go:334] "Generic (PLEG): container finished" podID="3270814b-f542-4706-bfaa-63403d036152" containerID="11d0f23969461ca00d3e586d504d52406021f5ad6cc06b657c3641de7da30974" exitCode=0 Oct 05 09:52:01 crc kubenswrapper[4935]: I1005 09:52:01.853726 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bxp77" event={"ID":"3270814b-f542-4706-bfaa-63403d036152","Type":"ContainerDied","Data":"11d0f23969461ca00d3e586d504d52406021f5ad6cc06b657c3641de7da30974"} Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.228964 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.346237 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-catalog-content\") pod \"3270814b-f542-4706-bfaa-63403d036152\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.346278 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtl5b\" (UniqueName: \"kubernetes.io/projected/3270814b-f542-4706-bfaa-63403d036152-kube-api-access-dtl5b\") pod \"3270814b-f542-4706-bfaa-63403d036152\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.346461 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-utilities\") pod \"3270814b-f542-4706-bfaa-63403d036152\" (UID: \"3270814b-f542-4706-bfaa-63403d036152\") " Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.347331 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-utilities" (OuterVolumeSpecName: "utilities") pod "3270814b-f542-4706-bfaa-63403d036152" (UID: "3270814b-f542-4706-bfaa-63403d036152"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.351989 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3270814b-f542-4706-bfaa-63403d036152-kube-api-access-dtl5b" (OuterVolumeSpecName: "kube-api-access-dtl5b") pod "3270814b-f542-4706-bfaa-63403d036152" (UID: "3270814b-f542-4706-bfaa-63403d036152"). InnerVolumeSpecName "kube-api-access-dtl5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.410561 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3270814b-f542-4706-bfaa-63403d036152" (UID: "3270814b-f542-4706-bfaa-63403d036152"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.448253 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.448285 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3270814b-f542-4706-bfaa-63403d036152-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.448297 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtl5b\" (UniqueName: \"kubernetes.io/projected/3270814b-f542-4706-bfaa-63403d036152-kube-api-access-dtl5b\") on node \"crc\" DevicePath \"\"" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.869032 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bxp77" event={"ID":"3270814b-f542-4706-bfaa-63403d036152","Type":"ContainerDied","Data":"0bd6300eec2cd6261d179014eabf45e91385902b7a11cc4ec0125104c0404992"} Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.869118 4935 scope.go:117] "RemoveContainer" containerID="11d0f23969461ca00d3e586d504d52406021f5ad6cc06b657c3641de7da30974" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.869143 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bxp77" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.904207 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bxp77"] Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.907369 4935 scope.go:117] "RemoveContainer" containerID="6997c33ea25080a5167f599c202eb35fd9e70440b9bfd51479ef9d8c77304f8d" Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.913296 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bxp77"] Oct 05 09:52:02 crc kubenswrapper[4935]: I1005 09:52:02.942019 4935 scope.go:117] "RemoveContainer" containerID="7c1d3754e0068018253e448791e276bb728cc5f4ec234e74500c5a1422707f21" Oct 05 09:52:04 crc kubenswrapper[4935]: I1005 09:52:04.791340 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3270814b-f542-4706-bfaa-63403d036152" path="/var/lib/kubelet/pods/3270814b-f542-4706-bfaa-63403d036152/volumes" Oct 05 09:53:44 crc kubenswrapper[4935]: I1005 09:53:44.289565 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:53:44 crc kubenswrapper[4935]: I1005 09:53:44.290304 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:54:14 crc kubenswrapper[4935]: I1005 09:54:14.289270 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:54:14 crc kubenswrapper[4935]: I1005 09:54:14.290745 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:54:44 crc kubenswrapper[4935]: I1005 09:54:44.290277 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:54:44 crc kubenswrapper[4935]: I1005 09:54:44.291796 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:54:44 crc kubenswrapper[4935]: I1005 09:54:44.292100 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:54:44 crc kubenswrapper[4935]: I1005 09:54:44.293633 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"58c70cc3730353e78569d63ee4affa370bf03d9815dc648f7d1464254f04339b"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:54:44 crc kubenswrapper[4935]: I1005 09:54:44.293784 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://58c70cc3730353e78569d63ee4affa370bf03d9815dc648f7d1464254f04339b" gracePeriod=600 Oct 05 09:54:45 crc kubenswrapper[4935]: I1005 09:54:45.173066 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="58c70cc3730353e78569d63ee4affa370bf03d9815dc648f7d1464254f04339b" exitCode=0 Oct 05 09:54:45 crc kubenswrapper[4935]: I1005 09:54:45.173120 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"58c70cc3730353e78569d63ee4affa370bf03d9815dc648f7d1464254f04339b"} Oct 05 09:54:45 crc kubenswrapper[4935]: I1005 09:54:45.174282 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78"} Oct 05 09:54:45 crc kubenswrapper[4935]: I1005 09:54:45.174328 4935 scope.go:117] "RemoveContainer" containerID="20110fbe5eefde5334e265a354b6695e055c505e8f2c6272f9fa69cf01d282d7" Oct 05 09:56:44 crc kubenswrapper[4935]: I1005 09:56:44.290443 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:56:44 crc kubenswrapper[4935]: I1005 09:56:44.290968 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:57:14 crc kubenswrapper[4935]: I1005 09:57:14.289842 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:57:14 crc kubenswrapper[4935]: I1005 09:57:14.290509 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.290036 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.290980 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.291052 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.291864 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.291973 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" gracePeriod=600 Oct 05 09:57:44 crc kubenswrapper[4935]: E1005 09:57:44.445083 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.748784 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" exitCode=0 Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.748850 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78"} Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.748924 4935 scope.go:117] "RemoveContainer" containerID="58c70cc3730353e78569d63ee4affa370bf03d9815dc648f7d1464254f04339b" Oct 05 09:57:44 crc kubenswrapper[4935]: I1005 09:57:44.749980 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:57:44 crc kubenswrapper[4935]: E1005 09:57:44.750358 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.559012 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pz5b2"] Oct 05 09:57:53 crc kubenswrapper[4935]: E1005 09:57:53.560342 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3270814b-f542-4706-bfaa-63403d036152" containerName="extract-content" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.560365 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3270814b-f542-4706-bfaa-63403d036152" containerName="extract-content" Oct 05 09:57:53 crc kubenswrapper[4935]: E1005 09:57:53.560393 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3270814b-f542-4706-bfaa-63403d036152" containerName="extract-utilities" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.560406 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3270814b-f542-4706-bfaa-63403d036152" containerName="extract-utilities" Oct 05 09:57:53 crc kubenswrapper[4935]: E1005 09:57:53.560451 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3270814b-f542-4706-bfaa-63403d036152" containerName="registry-server" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.560464 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="3270814b-f542-4706-bfaa-63403d036152" containerName="registry-server" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.560842 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="3270814b-f542-4706-bfaa-63403d036152" containerName="registry-server" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.563785 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.588738 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pz5b2"] Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.634788 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-catalog-content\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.635146 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csmmc\" (UniqueName: \"kubernetes.io/projected/acbbc06d-2c0c-4d05-b461-138d23dbc33e-kube-api-access-csmmc\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.635285 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-utilities\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.736806 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-utilities\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.736955 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-catalog-content\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.737020 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csmmc\" (UniqueName: \"kubernetes.io/projected/acbbc06d-2c0c-4d05-b461-138d23dbc33e-kube-api-access-csmmc\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.737656 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-utilities\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.737690 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-catalog-content\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.759127 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csmmc\" (UniqueName: \"kubernetes.io/projected/acbbc06d-2c0c-4d05-b461-138d23dbc33e-kube-api-access-csmmc\") pod \"redhat-operators-pz5b2\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:53 crc kubenswrapper[4935]: I1005 09:57:53.910552 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:57:54 crc kubenswrapper[4935]: I1005 09:57:54.422439 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pz5b2"] Oct 05 09:57:54 crc kubenswrapper[4935]: I1005 09:57:54.894171 4935 generic.go:334] "Generic (PLEG): container finished" podID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerID="d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947" exitCode=0 Oct 05 09:57:54 crc kubenswrapper[4935]: I1005 09:57:54.894211 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz5b2" event={"ID":"acbbc06d-2c0c-4d05-b461-138d23dbc33e","Type":"ContainerDied","Data":"d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947"} Oct 05 09:57:54 crc kubenswrapper[4935]: I1005 09:57:54.894519 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz5b2" event={"ID":"acbbc06d-2c0c-4d05-b461-138d23dbc33e","Type":"ContainerStarted","Data":"4574fb8c02706f9a697c5148975512a51e1511835a73cd9462b97ee8fef47b9b"} Oct 05 09:57:54 crc kubenswrapper[4935]: I1005 09:57:54.896143 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:57:55 crc kubenswrapper[4935]: I1005 09:57:55.906074 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz5b2" event={"ID":"acbbc06d-2c0c-4d05-b461-138d23dbc33e","Type":"ContainerStarted","Data":"bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b"} Oct 05 09:57:57 crc kubenswrapper[4935]: I1005 09:57:57.930105 4935 generic.go:334] "Generic (PLEG): container finished" podID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerID="bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b" exitCode=0 Oct 05 09:57:57 crc kubenswrapper[4935]: I1005 09:57:57.930487 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz5b2" event={"ID":"acbbc06d-2c0c-4d05-b461-138d23dbc33e","Type":"ContainerDied","Data":"bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b"} Oct 05 09:57:58 crc kubenswrapper[4935]: I1005 09:57:58.945768 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz5b2" event={"ID":"acbbc06d-2c0c-4d05-b461-138d23dbc33e","Type":"ContainerStarted","Data":"a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728"} Oct 05 09:57:58 crc kubenswrapper[4935]: I1005 09:57:58.986028 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pz5b2" podStartSLOduration=2.548458657 podStartE2EDuration="5.986003233s" podCreationTimestamp="2025-10-05 09:57:53 +0000 UTC" firstStartedPulling="2025-10-05 09:57:54.895871789 +0000 UTC m=+11108.778498239" lastFinishedPulling="2025-10-05 09:57:58.333416355 +0000 UTC m=+11112.216042815" observedRunningTime="2025-10-05 09:57:58.972344131 +0000 UTC m=+11112.854970591" watchObservedRunningTime="2025-10-05 09:57:58.986003233 +0000 UTC m=+11112.868629703" Oct 05 09:57:59 crc kubenswrapper[4935]: I1005 09:57:59.781155 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:57:59 crc kubenswrapper[4935]: E1005 09:57:59.781470 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:58:03 crc kubenswrapper[4935]: I1005 09:58:03.911710 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:58:03 crc kubenswrapper[4935]: I1005 09:58:03.912367 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:58:05 crc kubenswrapper[4935]: I1005 09:58:05.002767 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pz5b2" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="registry-server" probeResult="failure" output=< Oct 05 09:58:05 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 09:58:05 crc kubenswrapper[4935]: > Oct 05 09:58:11 crc kubenswrapper[4935]: I1005 09:58:11.778001 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:58:11 crc kubenswrapper[4935]: E1005 09:58:11.780090 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:58:13 crc kubenswrapper[4935]: I1005 09:58:13.981627 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:58:14 crc kubenswrapper[4935]: I1005 09:58:14.032402 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:58:14 crc kubenswrapper[4935]: I1005 09:58:14.222537 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pz5b2"] Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.155039 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pz5b2" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="registry-server" containerID="cri-o://a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728" gracePeriod=2 Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.647858 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.793515 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-catalog-content\") pod \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.795274 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-utilities\") pod \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.795826 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csmmc\" (UniqueName: \"kubernetes.io/projected/acbbc06d-2c0c-4d05-b461-138d23dbc33e-kube-api-access-csmmc\") pod \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\" (UID: \"acbbc06d-2c0c-4d05-b461-138d23dbc33e\") " Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.796396 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-utilities" (OuterVolumeSpecName: "utilities") pod "acbbc06d-2c0c-4d05-b461-138d23dbc33e" (UID: "acbbc06d-2c0c-4d05-b461-138d23dbc33e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.797355 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.808381 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acbbc06d-2c0c-4d05-b461-138d23dbc33e-kube-api-access-csmmc" (OuterVolumeSpecName: "kube-api-access-csmmc") pod "acbbc06d-2c0c-4d05-b461-138d23dbc33e" (UID: "acbbc06d-2c0c-4d05-b461-138d23dbc33e"). InnerVolumeSpecName "kube-api-access-csmmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.889857 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "acbbc06d-2c0c-4d05-b461-138d23dbc33e" (UID: "acbbc06d-2c0c-4d05-b461-138d23dbc33e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.900117 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csmmc\" (UniqueName: \"kubernetes.io/projected/acbbc06d-2c0c-4d05-b461-138d23dbc33e-kube-api-access-csmmc\") on node \"crc\" DevicePath \"\"" Oct 05 09:58:15 crc kubenswrapper[4935]: I1005 09:58:15.900178 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbc06d-2c0c-4d05-b461-138d23dbc33e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.175134 4935 generic.go:334] "Generic (PLEG): container finished" podID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerID="a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728" exitCode=0 Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.175213 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz5b2" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.175242 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz5b2" event={"ID":"acbbc06d-2c0c-4d05-b461-138d23dbc33e","Type":"ContainerDied","Data":"a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728"} Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.175770 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz5b2" event={"ID":"acbbc06d-2c0c-4d05-b461-138d23dbc33e","Type":"ContainerDied","Data":"4574fb8c02706f9a697c5148975512a51e1511835a73cd9462b97ee8fef47b9b"} Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.175811 4935 scope.go:117] "RemoveContainer" containerID="a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.233207 4935 scope.go:117] "RemoveContainer" containerID="bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.255342 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pz5b2"] Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.269872 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pz5b2"] Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.300359 4935 scope.go:117] "RemoveContainer" containerID="d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.335420 4935 scope.go:117] "RemoveContainer" containerID="a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728" Oct 05 09:58:16 crc kubenswrapper[4935]: E1005 09:58:16.336132 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728\": container with ID starting with a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728 not found: ID does not exist" containerID="a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.336201 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728"} err="failed to get container status \"a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728\": rpc error: code = NotFound desc = could not find container \"a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728\": container with ID starting with a1b8ba55192ba06e7fc0458abcfe71e81a6a1837658d4c0545eab29129ee2728 not found: ID does not exist" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.336240 4935 scope.go:117] "RemoveContainer" containerID="bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b" Oct 05 09:58:16 crc kubenswrapper[4935]: E1005 09:58:16.336710 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b\": container with ID starting with bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b not found: ID does not exist" containerID="bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.336749 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b"} err="failed to get container status \"bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b\": rpc error: code = NotFound desc = could not find container \"bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b\": container with ID starting with bea2d0ee21b3aefd8b83a91ce7b9272889a7c525636e2f08b1f1f302d55e048b not found: ID does not exist" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.336783 4935 scope.go:117] "RemoveContainer" containerID="d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947" Oct 05 09:58:16 crc kubenswrapper[4935]: E1005 09:58:16.337188 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947\": container with ID starting with d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947 not found: ID does not exist" containerID="d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.337264 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947"} err="failed to get container status \"d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947\": rpc error: code = NotFound desc = could not find container \"d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947\": container with ID starting with d0472c090b945d6ca8696708f2f15d0c85ba51488260aa041dc023b206589947 not found: ID does not exist" Oct 05 09:58:16 crc kubenswrapper[4935]: I1005 09:58:16.794688 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" path="/var/lib/kubelet/pods/acbbc06d-2c0c-4d05-b461-138d23dbc33e/volumes" Oct 05 09:58:23 crc kubenswrapper[4935]: I1005 09:58:23.778018 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:58:23 crc kubenswrapper[4935]: E1005 09:58:23.778998 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:58:37 crc kubenswrapper[4935]: I1005 09:58:37.778383 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:58:37 crc kubenswrapper[4935]: E1005 09:58:37.779798 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:58:52 crc kubenswrapper[4935]: I1005 09:58:52.778861 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:58:52 crc kubenswrapper[4935]: E1005 09:58:52.779990 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:59:07 crc kubenswrapper[4935]: I1005 09:59:07.778414 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:59:07 crc kubenswrapper[4935]: E1005 09:59:07.779638 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:59:18 crc kubenswrapper[4935]: I1005 09:59:18.778394 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:59:18 crc kubenswrapper[4935]: E1005 09:59:18.779753 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.239068 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 09:59:27 crc kubenswrapper[4935]: E1005 09:59:27.240669 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="registry-server" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.240697 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="registry-server" Oct 05 09:59:27 crc kubenswrapper[4935]: E1005 09:59:27.240769 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="extract-content" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.240783 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="extract-content" Oct 05 09:59:27 crc kubenswrapper[4935]: E1005 09:59:27.240826 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="extract-utilities" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.240840 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="extract-utilities" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.241325 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="acbbc06d-2c0c-4d05-b461-138d23dbc33e" containerName="registry-server" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.242972 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.257264 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.264808 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.266578 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.266831 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.267285 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sw94x" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.436628 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.436997 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.437048 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.437096 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.437141 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.437173 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.437254 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.437322 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt7r7\" (UniqueName: \"kubernetes.io/projected/c54822d6-7348-4d81-8681-f338af0badc1-kube-api-access-pt7r7\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.437378 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-config-data\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539356 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539406 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539440 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539469 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539499 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539524 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539555 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539599 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt7r7\" (UniqueName: \"kubernetes.io/projected/c54822d6-7348-4d81-8681-f338af0badc1-kube-api-access-pt7r7\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.539630 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-config-data\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.540643 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.540857 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-config-data\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.541279 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.541374 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.541450 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.549683 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.550330 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.551387 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.575049 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt7r7\" (UniqueName: \"kubernetes.io/projected/c54822d6-7348-4d81-8681-f338af0badc1-kube-api-access-pt7r7\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.603188 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " pod="openstack/tempest-tests-tempest" Oct 05 09:59:27 crc kubenswrapper[4935]: I1005 09:59:27.880680 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 09:59:28 crc kubenswrapper[4935]: I1005 09:59:28.448268 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 09:59:29 crc kubenswrapper[4935]: I1005 09:59:29.163413 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c54822d6-7348-4d81-8681-f338af0badc1","Type":"ContainerStarted","Data":"c75ffa72d41ef6b986ee5c85f13c264e919cffbccbd465537b2c6e86d1e7714a"} Oct 05 09:59:30 crc kubenswrapper[4935]: I1005 09:59:30.777841 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:59:30 crc kubenswrapper[4935]: E1005 09:59:30.778718 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:59:43 crc kubenswrapper[4935]: I1005 09:59:43.778487 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:59:43 crc kubenswrapper[4935]: E1005 09:59:43.779505 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 09:59:55 crc kubenswrapper[4935]: I1005 09:59:55.778007 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 09:59:55 crc kubenswrapper[4935]: E1005 09:59:55.779139 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.159137 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf"] Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.161128 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.164680 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.164705 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.190624 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf"] Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.196072 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29jj2\" (UniqueName: \"kubernetes.io/projected/841d8469-6c87-4fb9-a01d-857c360d9e23-kube-api-access-29jj2\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.196148 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841d8469-6c87-4fb9-a01d-857c360d9e23-config-volume\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.196404 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841d8469-6c87-4fb9-a01d-857c360d9e23-secret-volume\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.298982 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29jj2\" (UniqueName: \"kubernetes.io/projected/841d8469-6c87-4fb9-a01d-857c360d9e23-kube-api-access-29jj2\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.299429 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841d8469-6c87-4fb9-a01d-857c360d9e23-config-volume\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.300320 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841d8469-6c87-4fb9-a01d-857c360d9e23-config-volume\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.300440 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841d8469-6c87-4fb9-a01d-857c360d9e23-secret-volume\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.308523 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841d8469-6c87-4fb9-a01d-857c360d9e23-secret-volume\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.315020 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29jj2\" (UniqueName: \"kubernetes.io/projected/841d8469-6c87-4fb9-a01d-857c360d9e23-kube-api-access-29jj2\") pod \"collect-profiles-29327640-vgklf\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:00 crc kubenswrapper[4935]: I1005 10:00:00.489510 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:09 crc kubenswrapper[4935]: I1005 10:00:09.777490 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:00:09 crc kubenswrapper[4935]: E1005 10:00:09.778559 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:00:10 crc kubenswrapper[4935]: E1005 10:00:10.115683 4935 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:88dc57612f447daadb492dcf3ad854ac" Oct 05 10:00:10 crc kubenswrapper[4935]: E1005 10:00:10.115791 4935 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:88dc57612f447daadb492dcf3ad854ac" Oct 05 10:00:10 crc kubenswrapper[4935]: E1005 10:00:10.117666 4935 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:88dc57612f447daadb492dcf3ad854ac,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pt7r7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(c54822d6-7348-4d81-8681-f338af0badc1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 10:00:10 crc kubenswrapper[4935]: E1005 10:00:10.119003 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="c54822d6-7348-4d81-8681-f338af0badc1" Oct 05 10:00:10 crc kubenswrapper[4935]: I1005 10:00:10.689764 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf"] Oct 05 10:00:10 crc kubenswrapper[4935]: W1005 10:00:10.702275 4935 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod841d8469_6c87_4fb9_a01d_857c360d9e23.slice/crio-0c33a5b466bc8c013d46fd172b59006bbb9a2f95888377542b21fb1498115972 WatchSource:0}: Error finding container 0c33a5b466bc8c013d46fd172b59006bbb9a2f95888377542b21fb1498115972: Status 404 returned error can't find the container with id 0c33a5b466bc8c013d46fd172b59006bbb9a2f95888377542b21fb1498115972 Oct 05 10:00:10 crc kubenswrapper[4935]: E1005 10:00:10.789451 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:88dc57612f447daadb492dcf3ad854ac\\\"\"" pod="openstack/tempest-tests-tempest" podUID="c54822d6-7348-4d81-8681-f338af0badc1" Oct 05 10:00:10 crc kubenswrapper[4935]: I1005 10:00:10.794881 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" event={"ID":"841d8469-6c87-4fb9-a01d-857c360d9e23","Type":"ContainerStarted","Data":"0c33a5b466bc8c013d46fd172b59006bbb9a2f95888377542b21fb1498115972"} Oct 05 10:00:11 crc kubenswrapper[4935]: I1005 10:00:11.805534 4935 generic.go:334] "Generic (PLEG): container finished" podID="841d8469-6c87-4fb9-a01d-857c360d9e23" containerID="1e0557cbfed9e09aa4bfb5294ac4a7bced185672b7f9d99ab78b600cf680269f" exitCode=0 Oct 05 10:00:11 crc kubenswrapper[4935]: I1005 10:00:11.805669 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" event={"ID":"841d8469-6c87-4fb9-a01d-857c360d9e23","Type":"ContainerDied","Data":"1e0557cbfed9e09aa4bfb5294ac4a7bced185672b7f9d99ab78b600cf680269f"} Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.217596 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.265994 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29jj2\" (UniqueName: \"kubernetes.io/projected/841d8469-6c87-4fb9-a01d-857c360d9e23-kube-api-access-29jj2\") pod \"841d8469-6c87-4fb9-a01d-857c360d9e23\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.266130 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841d8469-6c87-4fb9-a01d-857c360d9e23-config-volume\") pod \"841d8469-6c87-4fb9-a01d-857c360d9e23\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.266392 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841d8469-6c87-4fb9-a01d-857c360d9e23-secret-volume\") pod \"841d8469-6c87-4fb9-a01d-857c360d9e23\" (UID: \"841d8469-6c87-4fb9-a01d-857c360d9e23\") " Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.267098 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/841d8469-6c87-4fb9-a01d-857c360d9e23-config-volume" (OuterVolumeSpecName: "config-volume") pod "841d8469-6c87-4fb9-a01d-857c360d9e23" (UID: "841d8469-6c87-4fb9-a01d-857c360d9e23"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.267619 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841d8469-6c87-4fb9-a01d-857c360d9e23-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.275706 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/841d8469-6c87-4fb9-a01d-857c360d9e23-kube-api-access-29jj2" (OuterVolumeSpecName: "kube-api-access-29jj2") pod "841d8469-6c87-4fb9-a01d-857c360d9e23" (UID: "841d8469-6c87-4fb9-a01d-857c360d9e23"). InnerVolumeSpecName "kube-api-access-29jj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.275863 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841d8469-6c87-4fb9-a01d-857c360d9e23-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "841d8469-6c87-4fb9-a01d-857c360d9e23" (UID: "841d8469-6c87-4fb9-a01d-857c360d9e23"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.370492 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841d8469-6c87-4fb9-a01d-857c360d9e23-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.370550 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29jj2\" (UniqueName: \"kubernetes.io/projected/841d8469-6c87-4fb9-a01d-857c360d9e23-kube-api-access-29jj2\") on node \"crc\" DevicePath \"\"" Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.835356 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" event={"ID":"841d8469-6c87-4fb9-a01d-857c360d9e23","Type":"ContainerDied","Data":"0c33a5b466bc8c013d46fd172b59006bbb9a2f95888377542b21fb1498115972"} Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.835408 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c33a5b466bc8c013d46fd172b59006bbb9a2f95888377542b21fb1498115972" Oct 05 10:00:13 crc kubenswrapper[4935]: I1005 10:00:13.835474 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327640-vgklf" Oct 05 10:00:14 crc kubenswrapper[4935]: I1005 10:00:14.322658 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4"] Oct 05 10:00:14 crc kubenswrapper[4935]: I1005 10:00:14.337694 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327595-rfhm4"] Oct 05 10:00:14 crc kubenswrapper[4935]: I1005 10:00:14.800056 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e72564f1-aaaf-4cdc-8221-2842954a46b7" path="/var/lib/kubelet/pods/e72564f1-aaaf-4cdc-8221-2842954a46b7/volumes" Oct 05 10:00:22 crc kubenswrapper[4935]: I1005 10:00:22.778608 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:00:22 crc kubenswrapper[4935]: E1005 10:00:22.779847 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:00:26 crc kubenswrapper[4935]: I1005 10:00:26.000310 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 05 10:00:28 crc kubenswrapper[4935]: I1005 10:00:28.027361 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c54822d6-7348-4d81-8681-f338af0badc1","Type":"ContainerStarted","Data":"75c1fa9680d767c40d2f9685ede1d068a8a5e0f33fad8d26fb898946eacedf54"} Oct 05 10:00:28 crc kubenswrapper[4935]: I1005 10:00:28.063788 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.52778146 podStartE2EDuration="1m2.063761407s" podCreationTimestamp="2025-10-05 09:59:26 +0000 UTC" firstStartedPulling="2025-10-05 09:59:28.46122795 +0000 UTC m=+11202.343854420" lastFinishedPulling="2025-10-05 10:00:25.997207907 +0000 UTC m=+11259.879834367" observedRunningTime="2025-10-05 10:00:28.052473758 +0000 UTC m=+11261.935100218" watchObservedRunningTime="2025-10-05 10:00:28.063761407 +0000 UTC m=+11261.946387877" Oct 05 10:00:35 crc kubenswrapper[4935]: I1005 10:00:35.778107 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:00:35 crc kubenswrapper[4935]: E1005 10:00:35.778992 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:00:47 crc kubenswrapper[4935]: I1005 10:00:47.778088 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:00:47 crc kubenswrapper[4935]: E1005 10:00:47.780036 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.169753 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29327641-kw7hz"] Oct 05 10:01:00 crc kubenswrapper[4935]: E1005 10:01:00.171134 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="841d8469-6c87-4fb9-a01d-857c360d9e23" containerName="collect-profiles" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.171153 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="841d8469-6c87-4fb9-a01d-857c360d9e23" containerName="collect-profiles" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.171470 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="841d8469-6c87-4fb9-a01d-857c360d9e23" containerName="collect-profiles" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.172364 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.213954 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29327641-kw7hz"] Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.279017 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-config-data\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.279075 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8znqz\" (UniqueName: \"kubernetes.io/projected/f004cad9-5506-4c21-8fbc-9bc72bb3470a-kube-api-access-8znqz\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.279150 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-fernet-keys\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.279331 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-combined-ca-bundle\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.381173 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-combined-ca-bundle\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.381340 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-config-data\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.381371 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8znqz\" (UniqueName: \"kubernetes.io/projected/f004cad9-5506-4c21-8fbc-9bc72bb3470a-kube-api-access-8znqz\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.381418 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-fernet-keys\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.390999 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-fernet-keys\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.394705 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-config-data\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.405805 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-combined-ca-bundle\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.414013 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8znqz\" (UniqueName: \"kubernetes.io/projected/f004cad9-5506-4c21-8fbc-9bc72bb3470a-kube-api-access-8znqz\") pod \"keystone-cron-29327641-kw7hz\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:00 crc kubenswrapper[4935]: I1005 10:01:00.497097 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:01 crc kubenswrapper[4935]: I1005 10:01:01.058743 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29327641-kw7hz"] Oct 05 10:01:01 crc kubenswrapper[4935]: I1005 10:01:01.510041 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327641-kw7hz" event={"ID":"f004cad9-5506-4c21-8fbc-9bc72bb3470a","Type":"ContainerStarted","Data":"53981f5bc99e018d42625219fde070682cc0c7bb9f9d8a90802daf51ca821e83"} Oct 05 10:01:01 crc kubenswrapper[4935]: I1005 10:01:01.510415 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327641-kw7hz" event={"ID":"f004cad9-5506-4c21-8fbc-9bc72bb3470a","Type":"ContainerStarted","Data":"330555cb4f47659db5844c0ea020d4e502ca231ef6ef070a5d6798b7b201a3dc"} Oct 05 10:01:01 crc kubenswrapper[4935]: I1005 10:01:01.529384 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29327641-kw7hz" podStartSLOduration=1.529324136 podStartE2EDuration="1.529324136s" podCreationTimestamp="2025-10-05 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 10:01:01.524167319 +0000 UTC m=+11295.406793779" watchObservedRunningTime="2025-10-05 10:01:01.529324136 +0000 UTC m=+11295.411950606" Oct 05 10:01:02 crc kubenswrapper[4935]: I1005 10:01:02.777938 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:01:02 crc kubenswrapper[4935]: E1005 10:01:02.778523 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:01:04 crc kubenswrapper[4935]: I1005 10:01:04.548707 4935 generic.go:334] "Generic (PLEG): container finished" podID="f004cad9-5506-4c21-8fbc-9bc72bb3470a" containerID="53981f5bc99e018d42625219fde070682cc0c7bb9f9d8a90802daf51ca821e83" exitCode=0 Oct 05 10:01:04 crc kubenswrapper[4935]: I1005 10:01:04.548830 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327641-kw7hz" event={"ID":"f004cad9-5506-4c21-8fbc-9bc72bb3470a","Type":"ContainerDied","Data":"53981f5bc99e018d42625219fde070682cc0c7bb9f9d8a90802daf51ca821e83"} Oct 05 10:01:05 crc kubenswrapper[4935]: I1005 10:01:05.473438 4935 scope.go:117] "RemoveContainer" containerID="94b5e9de84b890ea9584cd274e0065fc43d37db076c0d92aaaf87bdf7a92eb5c" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.019717 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.171936 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8znqz\" (UniqueName: \"kubernetes.io/projected/f004cad9-5506-4c21-8fbc-9bc72bb3470a-kube-api-access-8znqz\") pod \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.172089 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-fernet-keys\") pod \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.172198 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-config-data\") pod \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.172389 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-combined-ca-bundle\") pod \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\" (UID: \"f004cad9-5506-4c21-8fbc-9bc72bb3470a\") " Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.196616 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f004cad9-5506-4c21-8fbc-9bc72bb3470a" (UID: "f004cad9-5506-4c21-8fbc-9bc72bb3470a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.199220 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f004cad9-5506-4c21-8fbc-9bc72bb3470a-kube-api-access-8znqz" (OuterVolumeSpecName: "kube-api-access-8znqz") pod "f004cad9-5506-4c21-8fbc-9bc72bb3470a" (UID: "f004cad9-5506-4c21-8fbc-9bc72bb3470a"). InnerVolumeSpecName "kube-api-access-8znqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.248033 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f004cad9-5506-4c21-8fbc-9bc72bb3470a" (UID: "f004cad9-5506-4c21-8fbc-9bc72bb3470a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.277242 4935 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.277269 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8znqz\" (UniqueName: \"kubernetes.io/projected/f004cad9-5506-4c21-8fbc-9bc72bb3470a-kube-api-access-8znqz\") on node \"crc\" DevicePath \"\"" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.277280 4935 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.281047 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-config-data" (OuterVolumeSpecName: "config-data") pod "f004cad9-5506-4c21-8fbc-9bc72bb3470a" (UID: "f004cad9-5506-4c21-8fbc-9bc72bb3470a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.380817 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f004cad9-5506-4c21-8fbc-9bc72bb3470a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.574037 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327641-kw7hz" event={"ID":"f004cad9-5506-4c21-8fbc-9bc72bb3470a","Type":"ContainerDied","Data":"330555cb4f47659db5844c0ea020d4e502ca231ef6ef070a5d6798b7b201a3dc"} Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.574079 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="330555cb4f47659db5844c0ea020d4e502ca231ef6ef070a5d6798b7b201a3dc" Oct 05 10:01:06 crc kubenswrapper[4935]: I1005 10:01:06.574097 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327641-kw7hz" Oct 05 10:01:14 crc kubenswrapper[4935]: I1005 10:01:14.782160 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:01:14 crc kubenswrapper[4935]: E1005 10:01:14.783189 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:01:29 crc kubenswrapper[4935]: I1005 10:01:29.777794 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:01:29 crc kubenswrapper[4935]: E1005 10:01:29.778707 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:01:44 crc kubenswrapper[4935]: I1005 10:01:44.778226 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:01:44 crc kubenswrapper[4935]: E1005 10:01:44.779166 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.105319 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pq7kr"] Oct 05 10:01:53 crc kubenswrapper[4935]: E1005 10:01:53.107054 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f004cad9-5506-4c21-8fbc-9bc72bb3470a" containerName="keystone-cron" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.107077 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="f004cad9-5506-4c21-8fbc-9bc72bb3470a" containerName="keystone-cron" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.107457 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="f004cad9-5506-4c21-8fbc-9bc72bb3470a" containerName="keystone-cron" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.110237 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.149743 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pq7kr"] Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.175771 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-utilities\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.176674 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-catalog-content\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.177294 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghb7b\" (UniqueName: \"kubernetes.io/projected/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-kube-api-access-ghb7b\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.279944 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghb7b\" (UniqueName: \"kubernetes.io/projected/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-kube-api-access-ghb7b\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.280023 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-utilities\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.280085 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-catalog-content\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.280594 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-utilities\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.284321 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-catalog-content\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.309646 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghb7b\" (UniqueName: \"kubernetes.io/projected/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-kube-api-access-ghb7b\") pod \"redhat-marketplace-pq7kr\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:53 crc kubenswrapper[4935]: I1005 10:01:53.457320 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:01:54 crc kubenswrapper[4935]: I1005 10:01:54.005279 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pq7kr"] Oct 05 10:01:54 crc kubenswrapper[4935]: I1005 10:01:54.203516 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pq7kr" event={"ID":"8f8ade4e-47ff-4eab-8dc6-1ed68593a548","Type":"ContainerStarted","Data":"dba9b7beeb2bef77216370f085a3062f906fdb0a0028abf9deb5097fe6657a47"} Oct 05 10:01:55 crc kubenswrapper[4935]: I1005 10:01:55.222963 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pq7kr" event={"ID":"8f8ade4e-47ff-4eab-8dc6-1ed68593a548","Type":"ContainerDied","Data":"520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93"} Oct 05 10:01:55 crc kubenswrapper[4935]: I1005 10:01:55.222772 4935 generic.go:334] "Generic (PLEG): container finished" podID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerID="520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93" exitCode=0 Oct 05 10:01:56 crc kubenswrapper[4935]: I1005 10:01:56.251728 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pq7kr" event={"ID":"8f8ade4e-47ff-4eab-8dc6-1ed68593a548","Type":"ContainerStarted","Data":"6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8"} Oct 05 10:01:57 crc kubenswrapper[4935]: I1005 10:01:57.277017 4935 generic.go:334] "Generic (PLEG): container finished" podID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerID="6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8" exitCode=0 Oct 05 10:01:57 crc kubenswrapper[4935]: I1005 10:01:57.277143 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pq7kr" event={"ID":"8f8ade4e-47ff-4eab-8dc6-1ed68593a548","Type":"ContainerDied","Data":"6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8"} Oct 05 10:01:58 crc kubenswrapper[4935]: I1005 10:01:58.297728 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pq7kr" event={"ID":"8f8ade4e-47ff-4eab-8dc6-1ed68593a548","Type":"ContainerStarted","Data":"8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6"} Oct 05 10:01:58 crc kubenswrapper[4935]: I1005 10:01:58.323880 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pq7kr" podStartSLOduration=2.875933251 podStartE2EDuration="5.323851262s" podCreationTimestamp="2025-10-05 10:01:53 +0000 UTC" firstStartedPulling="2025-10-05 10:01:55.226180419 +0000 UTC m=+11349.108806889" lastFinishedPulling="2025-10-05 10:01:57.67409845 +0000 UTC m=+11351.556724900" observedRunningTime="2025-10-05 10:01:58.315007358 +0000 UTC m=+11352.197633818" watchObservedRunningTime="2025-10-05 10:01:58.323851262 +0000 UTC m=+11352.206477732" Oct 05 10:01:59 crc kubenswrapper[4935]: I1005 10:01:59.778093 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:01:59 crc kubenswrapper[4935]: E1005 10:01:59.778715 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:02:03 crc kubenswrapper[4935]: I1005 10:02:03.457688 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:02:03 crc kubenswrapper[4935]: I1005 10:02:03.458528 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:02:03 crc kubenswrapper[4935]: I1005 10:02:03.526800 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:02:04 crc kubenswrapper[4935]: I1005 10:02:04.436884 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:02:04 crc kubenswrapper[4935]: I1005 10:02:04.499830 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pq7kr"] Oct 05 10:02:06 crc kubenswrapper[4935]: I1005 10:02:06.397644 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pq7kr" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerName="registry-server" containerID="cri-o://8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6" gracePeriod=2 Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.017176 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.068929 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghb7b\" (UniqueName: \"kubernetes.io/projected/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-kube-api-access-ghb7b\") pod \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.068995 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-catalog-content\") pod \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.069131 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-utilities\") pod \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\" (UID: \"8f8ade4e-47ff-4eab-8dc6-1ed68593a548\") " Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.070684 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-utilities" (OuterVolumeSpecName: "utilities") pod "8f8ade4e-47ff-4eab-8dc6-1ed68593a548" (UID: "8f8ade4e-47ff-4eab-8dc6-1ed68593a548"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.080093 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-kube-api-access-ghb7b" (OuterVolumeSpecName: "kube-api-access-ghb7b") pod "8f8ade4e-47ff-4eab-8dc6-1ed68593a548" (UID: "8f8ade4e-47ff-4eab-8dc6-1ed68593a548"). InnerVolumeSpecName "kube-api-access-ghb7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.094759 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f8ade4e-47ff-4eab-8dc6-1ed68593a548" (UID: "8f8ade4e-47ff-4eab-8dc6-1ed68593a548"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.171310 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghb7b\" (UniqueName: \"kubernetes.io/projected/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-kube-api-access-ghb7b\") on node \"crc\" DevicePath \"\"" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.171353 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.171362 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8ade4e-47ff-4eab-8dc6-1ed68593a548-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.425575 4935 generic.go:334] "Generic (PLEG): container finished" podID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerID="8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6" exitCode=0 Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.425651 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pq7kr" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.425700 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pq7kr" event={"ID":"8f8ade4e-47ff-4eab-8dc6-1ed68593a548","Type":"ContainerDied","Data":"8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6"} Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.427418 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pq7kr" event={"ID":"8f8ade4e-47ff-4eab-8dc6-1ed68593a548","Type":"ContainerDied","Data":"dba9b7beeb2bef77216370f085a3062f906fdb0a0028abf9deb5097fe6657a47"} Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.427456 4935 scope.go:117] "RemoveContainer" containerID="8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.451884 4935 scope.go:117] "RemoveContainer" containerID="6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.479765 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pq7kr"] Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.504749 4935 scope.go:117] "RemoveContainer" containerID="520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.538250 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pq7kr"] Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.565064 4935 scope.go:117] "RemoveContainer" containerID="8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6" Oct 05 10:02:07 crc kubenswrapper[4935]: E1005 10:02:07.568018 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6\": container with ID starting with 8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6 not found: ID does not exist" containerID="8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.568076 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6"} err="failed to get container status \"8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6\": rpc error: code = NotFound desc = could not find container \"8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6\": container with ID starting with 8b54eb37a2f5b98ab31d0808f7704dff0463ac2c948de7d9cf4353b1e23adbf6 not found: ID does not exist" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.568109 4935 scope.go:117] "RemoveContainer" containerID="6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8" Oct 05 10:02:07 crc kubenswrapper[4935]: E1005 10:02:07.568565 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8\": container with ID starting with 6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8 not found: ID does not exist" containerID="6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.568623 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8"} err="failed to get container status \"6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8\": rpc error: code = NotFound desc = could not find container \"6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8\": container with ID starting with 6bd6a84acd575376f2da34c341e2fe3a233f5ca6db72d435645bd7cab6909cf8 not found: ID does not exist" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.568666 4935 scope.go:117] "RemoveContainer" containerID="520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93" Oct 05 10:02:07 crc kubenswrapper[4935]: E1005 10:02:07.569049 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93\": container with ID starting with 520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93 not found: ID does not exist" containerID="520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93" Oct 05 10:02:07 crc kubenswrapper[4935]: I1005 10:02:07.569074 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93"} err="failed to get container status \"520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93\": rpc error: code = NotFound desc = could not find container \"520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93\": container with ID starting with 520f91e7efd201d34d73083654d196368a6586a1db1bebb560b00ab86c3a9d93 not found: ID does not exist" Oct 05 10:02:08 crc kubenswrapper[4935]: I1005 10:02:08.799911 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" path="/var/lib/kubelet/pods/8f8ade4e-47ff-4eab-8dc6-1ed68593a548/volumes" Oct 05 10:02:13 crc kubenswrapper[4935]: I1005 10:02:13.778553 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:02:13 crc kubenswrapper[4935]: E1005 10:02:13.779503 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:02:28 crc kubenswrapper[4935]: I1005 10:02:28.777729 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:02:28 crc kubenswrapper[4935]: E1005 10:02:28.778464 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:02:39 crc kubenswrapper[4935]: I1005 10:02:39.777452 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:02:39 crc kubenswrapper[4935]: E1005 10:02:39.778212 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:02:51 crc kubenswrapper[4935]: I1005 10:02:51.777191 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:02:52 crc kubenswrapper[4935]: I1005 10:02:52.981986 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"e52f5c49d62f14761e6421fc5a28e06ecceb1b3129fff97230ea91fb748ee045"} Oct 05 10:05:14 crc kubenswrapper[4935]: I1005 10:05:14.289514 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:05:14 crc kubenswrapper[4935]: I1005 10:05:14.290090 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:05:44 crc kubenswrapper[4935]: I1005 10:05:44.289407 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:05:44 crc kubenswrapper[4935]: I1005 10:05:44.290086 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:06:14 crc kubenswrapper[4935]: I1005 10:06:14.289871 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:06:14 crc kubenswrapper[4935]: I1005 10:06:14.291247 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:06:14 crc kubenswrapper[4935]: I1005 10:06:14.291352 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 10:06:14 crc kubenswrapper[4935]: I1005 10:06:14.293308 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e52f5c49d62f14761e6421fc5a28e06ecceb1b3129fff97230ea91fb748ee045"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 10:06:14 crc kubenswrapper[4935]: I1005 10:06:14.293458 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://e52f5c49d62f14761e6421fc5a28e06ecceb1b3129fff97230ea91fb748ee045" gracePeriod=600 Oct 05 10:06:14 crc kubenswrapper[4935]: I1005 10:06:14.466548 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="e52f5c49d62f14761e6421fc5a28e06ecceb1b3129fff97230ea91fb748ee045" exitCode=0 Oct 05 10:06:14 crc kubenswrapper[4935]: I1005 10:06:14.466604 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"e52f5c49d62f14761e6421fc5a28e06ecceb1b3129fff97230ea91fb748ee045"} Oct 05 10:06:14 crc kubenswrapper[4935]: I1005 10:06:14.466649 4935 scope.go:117] "RemoveContainer" containerID="d510e250d47228087628c626604741b2d6b0de6b91797d981497a3b8f2900e78" Oct 05 10:06:15 crc kubenswrapper[4935]: I1005 10:06:15.480752 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4"} Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.346528 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5hrpx"] Oct 05 10:06:30 crc kubenswrapper[4935]: E1005 10:06:30.348167 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerName="registry-server" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.348202 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerName="registry-server" Oct 05 10:06:30 crc kubenswrapper[4935]: E1005 10:06:30.348216 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerName="extract-content" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.348223 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerName="extract-content" Oct 05 10:06:30 crc kubenswrapper[4935]: E1005 10:06:30.348264 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerName="extract-utilities" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.348271 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerName="extract-utilities" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.348561 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8ade4e-47ff-4eab-8dc6-1ed68593a548" containerName="registry-server" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.350159 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.370303 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5hrpx"] Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.458039 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-catalog-content\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.458258 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-utilities\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.458508 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk84f\" (UniqueName: \"kubernetes.io/projected/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-kube-api-access-gk84f\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.546766 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hwth8"] Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.549363 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.566043 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-catalog-content\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.566211 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-utilities\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.566325 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk84f\" (UniqueName: \"kubernetes.io/projected/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-kube-api-access-gk84f\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.566836 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-catalog-content\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.566933 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-utilities\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.583448 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hwth8"] Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.593787 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk84f\" (UniqueName: \"kubernetes.io/projected/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-kube-api-access-gk84f\") pod \"certified-operators-5hrpx\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.669793 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-utilities\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.669941 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-catalog-content\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.670018 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhqmr\" (UniqueName: \"kubernetes.io/projected/e7d85a89-3d12-4e91-ab27-a18e2098daff-kube-api-access-nhqmr\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.673743 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.772363 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-utilities\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.772807 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-catalog-content\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.772853 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhqmr\" (UniqueName: \"kubernetes.io/projected/e7d85a89-3d12-4e91-ab27-a18e2098daff-kube-api-access-nhqmr\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.773615 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-utilities\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.773646 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-catalog-content\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.802661 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhqmr\" (UniqueName: \"kubernetes.io/projected/e7d85a89-3d12-4e91-ab27-a18e2098daff-kube-api-access-nhqmr\") pod \"community-operators-hwth8\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:30 crc kubenswrapper[4935]: I1005 10:06:30.892818 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:31 crc kubenswrapper[4935]: I1005 10:06:31.285258 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5hrpx"] Oct 05 10:06:31 crc kubenswrapper[4935]: I1005 10:06:31.496048 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hwth8"] Oct 05 10:06:31 crc kubenswrapper[4935]: I1005 10:06:31.661541 4935 generic.go:334] "Generic (PLEG): container finished" podID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerID="aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342" exitCode=0 Oct 05 10:06:31 crc kubenswrapper[4935]: I1005 10:06:31.661724 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hrpx" event={"ID":"47ff89b9-8dcc-4f24-9df5-6d2b011e1552","Type":"ContainerDied","Data":"aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342"} Oct 05 10:06:31 crc kubenswrapper[4935]: I1005 10:06:31.661951 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hrpx" event={"ID":"47ff89b9-8dcc-4f24-9df5-6d2b011e1552","Type":"ContainerStarted","Data":"53d41ac548bf4c85fea94eda35e402a14e083bb040dedacc96b92739eb1583a4"} Oct 05 10:06:31 crc kubenswrapper[4935]: I1005 10:06:31.663729 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 10:06:31 crc kubenswrapper[4935]: I1005 10:06:31.664879 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hwth8" event={"ID":"e7d85a89-3d12-4e91-ab27-a18e2098daff","Type":"ContainerStarted","Data":"ace2919f27cb7896d262741d67201815773c15fac0a750fbf6d3b96c7184e33b"} Oct 05 10:06:32 crc kubenswrapper[4935]: I1005 10:06:32.698312 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hrpx" event={"ID":"47ff89b9-8dcc-4f24-9df5-6d2b011e1552","Type":"ContainerStarted","Data":"e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f"} Oct 05 10:06:32 crc kubenswrapper[4935]: I1005 10:06:32.702448 4935 generic.go:334] "Generic (PLEG): container finished" podID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerID="c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c" exitCode=0 Oct 05 10:06:32 crc kubenswrapper[4935]: I1005 10:06:32.702500 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hwth8" event={"ID":"e7d85a89-3d12-4e91-ab27-a18e2098daff","Type":"ContainerDied","Data":"c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c"} Oct 05 10:06:34 crc kubenswrapper[4935]: I1005 10:06:34.761540 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hwth8" event={"ID":"e7d85a89-3d12-4e91-ab27-a18e2098daff","Type":"ContainerStarted","Data":"6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a"} Oct 05 10:06:34 crc kubenswrapper[4935]: I1005 10:06:34.769824 4935 generic.go:334] "Generic (PLEG): container finished" podID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerID="e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f" exitCode=0 Oct 05 10:06:34 crc kubenswrapper[4935]: I1005 10:06:34.769881 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hrpx" event={"ID":"47ff89b9-8dcc-4f24-9df5-6d2b011e1552","Type":"ContainerDied","Data":"e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f"} Oct 05 10:06:35 crc kubenswrapper[4935]: I1005 10:06:35.784650 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hrpx" event={"ID":"47ff89b9-8dcc-4f24-9df5-6d2b011e1552","Type":"ContainerStarted","Data":"16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9"} Oct 05 10:06:35 crc kubenswrapper[4935]: I1005 10:06:35.789870 4935 generic.go:334] "Generic (PLEG): container finished" podID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerID="6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a" exitCode=0 Oct 05 10:06:35 crc kubenswrapper[4935]: I1005 10:06:35.789939 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hwth8" event={"ID":"e7d85a89-3d12-4e91-ab27-a18e2098daff","Type":"ContainerDied","Data":"6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a"} Oct 05 10:06:35 crc kubenswrapper[4935]: I1005 10:06:35.827831 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5hrpx" podStartSLOduration=2.115043008 podStartE2EDuration="5.827802867s" podCreationTimestamp="2025-10-05 10:06:30 +0000 UTC" firstStartedPulling="2025-10-05 10:06:31.663515661 +0000 UTC m=+11625.546142121" lastFinishedPulling="2025-10-05 10:06:35.37627552 +0000 UTC m=+11629.258901980" observedRunningTime="2025-10-05 10:06:35.817283019 +0000 UTC m=+11629.699909489" watchObservedRunningTime="2025-10-05 10:06:35.827802867 +0000 UTC m=+11629.710429337" Oct 05 10:06:36 crc kubenswrapper[4935]: I1005 10:06:36.809506 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hwth8" event={"ID":"e7d85a89-3d12-4e91-ab27-a18e2098daff","Type":"ContainerStarted","Data":"700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6"} Oct 05 10:06:36 crc kubenswrapper[4935]: I1005 10:06:36.842581 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hwth8" podStartSLOduration=3.378377226 podStartE2EDuration="6.842555022s" podCreationTimestamp="2025-10-05 10:06:30 +0000 UTC" firstStartedPulling="2025-10-05 10:06:32.708113811 +0000 UTC m=+11626.590740271" lastFinishedPulling="2025-10-05 10:06:36.172291607 +0000 UTC m=+11630.054918067" observedRunningTime="2025-10-05 10:06:36.838395162 +0000 UTC m=+11630.721021642" watchObservedRunningTime="2025-10-05 10:06:36.842555022 +0000 UTC m=+11630.725181482" Oct 05 10:06:40 crc kubenswrapper[4935]: I1005 10:06:40.674977 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:40 crc kubenswrapper[4935]: I1005 10:06:40.675469 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:40 crc kubenswrapper[4935]: I1005 10:06:40.894025 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:40 crc kubenswrapper[4935]: I1005 10:06:40.895307 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:40 crc kubenswrapper[4935]: I1005 10:06:40.948672 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:41 crc kubenswrapper[4935]: I1005 10:06:41.734263 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-5hrpx" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="registry-server" probeResult="failure" output=< Oct 05 10:06:41 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 10:06:41 crc kubenswrapper[4935]: > Oct 05 10:06:41 crc kubenswrapper[4935]: I1005 10:06:41.907507 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:44 crc kubenswrapper[4935]: I1005 10:06:44.136961 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hwth8"] Oct 05 10:06:44 crc kubenswrapper[4935]: I1005 10:06:44.896529 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hwth8" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerName="registry-server" containerID="cri-o://700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6" gracePeriod=2 Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.703579 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.832820 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-catalog-content\") pod \"e7d85a89-3d12-4e91-ab27-a18e2098daff\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.833204 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhqmr\" (UniqueName: \"kubernetes.io/projected/e7d85a89-3d12-4e91-ab27-a18e2098daff-kube-api-access-nhqmr\") pod \"e7d85a89-3d12-4e91-ab27-a18e2098daff\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.833327 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-utilities\") pod \"e7d85a89-3d12-4e91-ab27-a18e2098daff\" (UID: \"e7d85a89-3d12-4e91-ab27-a18e2098daff\") " Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.833922 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-utilities" (OuterVolumeSpecName: "utilities") pod "e7d85a89-3d12-4e91-ab27-a18e2098daff" (UID: "e7d85a89-3d12-4e91-ab27-a18e2098daff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.839545 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7d85a89-3d12-4e91-ab27-a18e2098daff-kube-api-access-nhqmr" (OuterVolumeSpecName: "kube-api-access-nhqmr") pod "e7d85a89-3d12-4e91-ab27-a18e2098daff" (UID: "e7d85a89-3d12-4e91-ab27-a18e2098daff"). InnerVolumeSpecName "kube-api-access-nhqmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.888548 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7d85a89-3d12-4e91-ab27-a18e2098daff" (UID: "e7d85a89-3d12-4e91-ab27-a18e2098daff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.910244 4935 generic.go:334] "Generic (PLEG): container finished" podID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerID="700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6" exitCode=0 Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.910291 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hwth8" event={"ID":"e7d85a89-3d12-4e91-ab27-a18e2098daff","Type":"ContainerDied","Data":"700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6"} Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.910320 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hwth8" event={"ID":"e7d85a89-3d12-4e91-ab27-a18e2098daff","Type":"ContainerDied","Data":"ace2919f27cb7896d262741d67201815773c15fac0a750fbf6d3b96c7184e33b"} Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.910336 4935 scope.go:117] "RemoveContainer" containerID="700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.910450 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hwth8" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.935665 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.935696 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d85a89-3d12-4e91-ab27-a18e2098daff-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.935733 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhqmr\" (UniqueName: \"kubernetes.io/projected/e7d85a89-3d12-4e91-ab27-a18e2098daff-kube-api-access-nhqmr\") on node \"crc\" DevicePath \"\"" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.936102 4935 scope.go:117] "RemoveContainer" containerID="6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a" Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.957928 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hwth8"] Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.967335 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hwth8"] Oct 05 10:06:45 crc kubenswrapper[4935]: I1005 10:06:45.976827 4935 scope.go:117] "RemoveContainer" containerID="c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c" Oct 05 10:06:46 crc kubenswrapper[4935]: I1005 10:06:46.018256 4935 scope.go:117] "RemoveContainer" containerID="700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6" Oct 05 10:06:46 crc kubenswrapper[4935]: E1005 10:06:46.018952 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6\": container with ID starting with 700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6 not found: ID does not exist" containerID="700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6" Oct 05 10:06:46 crc kubenswrapper[4935]: I1005 10:06:46.019001 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6"} err="failed to get container status \"700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6\": rpc error: code = NotFound desc = could not find container \"700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6\": container with ID starting with 700e3338c6c21f341b38f8d95b885f1f2fa223c424a2f3645c89cdf1334c10d6 not found: ID does not exist" Oct 05 10:06:46 crc kubenswrapper[4935]: I1005 10:06:46.019032 4935 scope.go:117] "RemoveContainer" containerID="6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a" Oct 05 10:06:46 crc kubenswrapper[4935]: E1005 10:06:46.019399 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a\": container with ID starting with 6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a not found: ID does not exist" containerID="6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a" Oct 05 10:06:46 crc kubenswrapper[4935]: I1005 10:06:46.019433 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a"} err="failed to get container status \"6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a\": rpc error: code = NotFound desc = could not find container \"6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a\": container with ID starting with 6bc2560dcd5f42092682a70739c234f4b23928334527a60508abcd04a023c37a not found: ID does not exist" Oct 05 10:06:46 crc kubenswrapper[4935]: I1005 10:06:46.019455 4935 scope.go:117] "RemoveContainer" containerID="c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c" Oct 05 10:06:46 crc kubenswrapper[4935]: E1005 10:06:46.019682 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c\": container with ID starting with c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c not found: ID does not exist" containerID="c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c" Oct 05 10:06:46 crc kubenswrapper[4935]: I1005 10:06:46.019709 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c"} err="failed to get container status \"c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c\": rpc error: code = NotFound desc = could not find container \"c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c\": container with ID starting with c746facb38667681364d4885ca5fef26688006a95fac4a9a522f4ccd92b96e1c not found: ID does not exist" Oct 05 10:06:46 crc kubenswrapper[4935]: I1005 10:06:46.792072 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" path="/var/lib/kubelet/pods/e7d85a89-3d12-4e91-ab27-a18e2098daff/volumes" Oct 05 10:06:50 crc kubenswrapper[4935]: I1005 10:06:50.745079 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:50 crc kubenswrapper[4935]: I1005 10:06:50.814152 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:50 crc kubenswrapper[4935]: I1005 10:06:50.993510 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5hrpx"] Oct 05 10:06:51 crc kubenswrapper[4935]: I1005 10:06:51.989790 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5hrpx" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="registry-server" containerID="cri-o://16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9" gracePeriod=2 Oct 05 10:06:52 crc kubenswrapper[4935]: I1005 10:06:52.781431 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:52 crc kubenswrapper[4935]: I1005 10:06:52.899313 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-catalog-content\") pod \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " Oct 05 10:06:52 crc kubenswrapper[4935]: I1005 10:06:52.899369 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk84f\" (UniqueName: \"kubernetes.io/projected/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-kube-api-access-gk84f\") pod \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " Oct 05 10:06:52 crc kubenswrapper[4935]: I1005 10:06:52.899459 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-utilities\") pod \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\" (UID: \"47ff89b9-8dcc-4f24-9df5-6d2b011e1552\") " Oct 05 10:06:52 crc kubenswrapper[4935]: I1005 10:06:52.900282 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-utilities" (OuterVolumeSpecName: "utilities") pod "47ff89b9-8dcc-4f24-9df5-6d2b011e1552" (UID: "47ff89b9-8dcc-4f24-9df5-6d2b011e1552"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:06:52 crc kubenswrapper[4935]: I1005 10:06:52.901585 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:06:52 crc kubenswrapper[4935]: I1005 10:06:52.907173 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-kube-api-access-gk84f" (OuterVolumeSpecName: "kube-api-access-gk84f") pod "47ff89b9-8dcc-4f24-9df5-6d2b011e1552" (UID: "47ff89b9-8dcc-4f24-9df5-6d2b011e1552"). InnerVolumeSpecName "kube-api-access-gk84f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:06:52 crc kubenswrapper[4935]: I1005 10:06:52.963057 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47ff89b9-8dcc-4f24-9df5-6d2b011e1552" (UID: "47ff89b9-8dcc-4f24-9df5-6d2b011e1552"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.002452 4935 generic.go:334] "Generic (PLEG): container finished" podID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerID="16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9" exitCode=0 Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.003023 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.003391 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk84f\" (UniqueName: \"kubernetes.io/projected/47ff89b9-8dcc-4f24-9df5-6d2b011e1552-kube-api-access-gk84f\") on node \"crc\" DevicePath \"\"" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.003112 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hrpx" event={"ID":"47ff89b9-8dcc-4f24-9df5-6d2b011e1552","Type":"ContainerDied","Data":"16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9"} Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.003530 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hrpx" event={"ID":"47ff89b9-8dcc-4f24-9df5-6d2b011e1552","Type":"ContainerDied","Data":"53d41ac548bf4c85fea94eda35e402a14e083bb040dedacc96b92739eb1583a4"} Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.003600 4935 scope.go:117] "RemoveContainer" containerID="16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.003098 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hrpx" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.037149 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5hrpx"] Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.039496 4935 scope.go:117] "RemoveContainer" containerID="e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.056964 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5hrpx"] Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.062766 4935 scope.go:117] "RemoveContainer" containerID="aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.109246 4935 scope.go:117] "RemoveContainer" containerID="16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9" Oct 05 10:06:53 crc kubenswrapper[4935]: E1005 10:06:53.109962 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9\": container with ID starting with 16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9 not found: ID does not exist" containerID="16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.110023 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9"} err="failed to get container status \"16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9\": rpc error: code = NotFound desc = could not find container \"16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9\": container with ID starting with 16d9042e602876b1c627f2c7d89111c2b703fe9ec612e1d53e57bc0cd756d0c9 not found: ID does not exist" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.110062 4935 scope.go:117] "RemoveContainer" containerID="e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f" Oct 05 10:06:53 crc kubenswrapper[4935]: E1005 10:06:53.110487 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f\": container with ID starting with e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f not found: ID does not exist" containerID="e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.110523 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f"} err="failed to get container status \"e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f\": rpc error: code = NotFound desc = could not find container \"e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f\": container with ID starting with e52b3604e7515b6befcaae047573c3d5c99cb4d01522e96f397a995c868b6c2f not found: ID does not exist" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.110569 4935 scope.go:117] "RemoveContainer" containerID="aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342" Oct 05 10:06:53 crc kubenswrapper[4935]: E1005 10:06:53.110853 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342\": container with ID starting with aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342 not found: ID does not exist" containerID="aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342" Oct 05 10:06:53 crc kubenswrapper[4935]: I1005 10:06:53.110933 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342"} err="failed to get container status \"aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342\": rpc error: code = NotFound desc = could not find container \"aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342\": container with ID starting with aef617025d9dc4d423789de3d62a72762f206b8e0dad42cca6788259b026f342 not found: ID does not exist" Oct 05 10:06:54 crc kubenswrapper[4935]: I1005 10:06:54.792659 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" path="/var/lib/kubelet/pods/47ff89b9-8dcc-4f24-9df5-6d2b011e1552/volumes" Oct 05 10:08:14 crc kubenswrapper[4935]: I1005 10:08:14.288919 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:08:14 crc kubenswrapper[4935]: I1005 10:08:14.289681 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:08:44 crc kubenswrapper[4935]: I1005 10:08:44.289124 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:08:44 crc kubenswrapper[4935]: I1005 10:08:44.290041 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.289467 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.290329 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.290383 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.291105 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.291174 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" gracePeriod=600 Oct 05 10:09:14 crc kubenswrapper[4935]: E1005 10:09:14.421351 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.638406 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" exitCode=0 Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.638484 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4"} Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.638584 4935 scope.go:117] "RemoveContainer" containerID="e52f5c49d62f14761e6421fc5a28e06ecceb1b3129fff97230ea91fb748ee045" Oct 05 10:09:14 crc kubenswrapper[4935]: I1005 10:09:14.640428 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:09:14 crc kubenswrapper[4935]: E1005 10:09:14.641017 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.831631 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k8lkx"] Oct 05 10:09:19 crc kubenswrapper[4935]: E1005 10:09:19.832882 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerName="extract-content" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.832919 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerName="extract-content" Oct 05 10:09:19 crc kubenswrapper[4935]: E1005 10:09:19.832929 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="extract-content" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.832935 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="extract-content" Oct 05 10:09:19 crc kubenswrapper[4935]: E1005 10:09:19.832956 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="registry-server" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.832965 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="registry-server" Oct 05 10:09:19 crc kubenswrapper[4935]: E1005 10:09:19.832989 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerName="extract-utilities" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.832998 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerName="extract-utilities" Oct 05 10:09:19 crc kubenswrapper[4935]: E1005 10:09:19.833013 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerName="registry-server" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.833021 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerName="registry-server" Oct 05 10:09:19 crc kubenswrapper[4935]: E1005 10:09:19.833048 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="extract-utilities" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.833057 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="extract-utilities" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.833286 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7d85a89-3d12-4e91-ab27-a18e2098daff" containerName="registry-server" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.833365 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ff89b9-8dcc-4f24-9df5-6d2b011e1552" containerName="registry-server" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.835187 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.849247 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k8lkx"] Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.899076 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-catalog-content\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.899152 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-utilities\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:19 crc kubenswrapper[4935]: I1005 10:09:19.899366 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxp5l\" (UniqueName: \"kubernetes.io/projected/7763943d-025a-47e6-9afc-b8b02d21c876-kube-api-access-lxp5l\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:20 crc kubenswrapper[4935]: I1005 10:09:20.002342 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-catalog-content\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:20 crc kubenswrapper[4935]: I1005 10:09:20.002669 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-utilities\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:20 crc kubenswrapper[4935]: I1005 10:09:20.002807 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxp5l\" (UniqueName: \"kubernetes.io/projected/7763943d-025a-47e6-9afc-b8b02d21c876-kube-api-access-lxp5l\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:20 crc kubenswrapper[4935]: I1005 10:09:20.003027 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-catalog-content\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:20 crc kubenswrapper[4935]: I1005 10:09:20.003466 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-utilities\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:20 crc kubenswrapper[4935]: I1005 10:09:20.025269 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxp5l\" (UniqueName: \"kubernetes.io/projected/7763943d-025a-47e6-9afc-b8b02d21c876-kube-api-access-lxp5l\") pod \"redhat-operators-k8lkx\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:20 crc kubenswrapper[4935]: I1005 10:09:20.163952 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:20 crc kubenswrapper[4935]: I1005 10:09:20.704377 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k8lkx"] Oct 05 10:09:21 crc kubenswrapper[4935]: I1005 10:09:21.731725 4935 generic.go:334] "Generic (PLEG): container finished" podID="7763943d-025a-47e6-9afc-b8b02d21c876" containerID="4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e" exitCode=0 Oct 05 10:09:21 crc kubenswrapper[4935]: I1005 10:09:21.732034 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lkx" event={"ID":"7763943d-025a-47e6-9afc-b8b02d21c876","Type":"ContainerDied","Data":"4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e"} Oct 05 10:09:21 crc kubenswrapper[4935]: I1005 10:09:21.732066 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lkx" event={"ID":"7763943d-025a-47e6-9afc-b8b02d21c876","Type":"ContainerStarted","Data":"7379c3969393b7ccb437891b6d8dcff1d0975749507c2e3017e97838a933e647"} Oct 05 10:09:23 crc kubenswrapper[4935]: I1005 10:09:23.764358 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lkx" event={"ID":"7763943d-025a-47e6-9afc-b8b02d21c876","Type":"ContainerStarted","Data":"a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09"} Oct 05 10:09:26 crc kubenswrapper[4935]: I1005 10:09:26.805005 4935 generic.go:334] "Generic (PLEG): container finished" podID="7763943d-025a-47e6-9afc-b8b02d21c876" containerID="a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09" exitCode=0 Oct 05 10:09:26 crc kubenswrapper[4935]: I1005 10:09:26.805236 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lkx" event={"ID":"7763943d-025a-47e6-9afc-b8b02d21c876","Type":"ContainerDied","Data":"a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09"} Oct 05 10:09:27 crc kubenswrapper[4935]: I1005 10:09:27.778239 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:09:27 crc kubenswrapper[4935]: E1005 10:09:27.778793 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:09:27 crc kubenswrapper[4935]: I1005 10:09:27.819928 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lkx" event={"ID":"7763943d-025a-47e6-9afc-b8b02d21c876","Type":"ContainerStarted","Data":"476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d"} Oct 05 10:09:27 crc kubenswrapper[4935]: I1005 10:09:27.845201 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k8lkx" podStartSLOduration=3.308254074 podStartE2EDuration="8.845179333s" podCreationTimestamp="2025-10-05 10:09:19 +0000 UTC" firstStartedPulling="2025-10-05 10:09:21.734455891 +0000 UTC m=+11795.617082351" lastFinishedPulling="2025-10-05 10:09:27.27138116 +0000 UTC m=+11801.154007610" observedRunningTime="2025-10-05 10:09:27.841594138 +0000 UTC m=+11801.724220618" watchObservedRunningTime="2025-10-05 10:09:27.845179333 +0000 UTC m=+11801.727805793" Oct 05 10:09:30 crc kubenswrapper[4935]: I1005 10:09:30.164473 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:30 crc kubenswrapper[4935]: I1005 10:09:30.165237 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:31 crc kubenswrapper[4935]: I1005 10:09:31.219394 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k8lkx" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="registry-server" probeResult="failure" output=< Oct 05 10:09:31 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 10:09:31 crc kubenswrapper[4935]: > Oct 05 10:09:40 crc kubenswrapper[4935]: I1005 10:09:40.236861 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:40 crc kubenswrapper[4935]: I1005 10:09:40.305876 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:40 crc kubenswrapper[4935]: I1005 10:09:40.484133 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k8lkx"] Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.003785 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k8lkx" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="registry-server" containerID="cri-o://476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d" gracePeriod=2 Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.735455 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.784951 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:09:42 crc kubenswrapper[4935]: E1005 10:09:42.786057 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.784300 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxp5l\" (UniqueName: \"kubernetes.io/projected/7763943d-025a-47e6-9afc-b8b02d21c876-kube-api-access-lxp5l\") pod \"7763943d-025a-47e6-9afc-b8b02d21c876\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.786588 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-catalog-content\") pod \"7763943d-025a-47e6-9afc-b8b02d21c876\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.786824 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-utilities\") pod \"7763943d-025a-47e6-9afc-b8b02d21c876\" (UID: \"7763943d-025a-47e6-9afc-b8b02d21c876\") " Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.790463 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-utilities" (OuterVolumeSpecName: "utilities") pod "7763943d-025a-47e6-9afc-b8b02d21c876" (UID: "7763943d-025a-47e6-9afc-b8b02d21c876"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.801148 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7763943d-025a-47e6-9afc-b8b02d21c876-kube-api-access-lxp5l" (OuterVolumeSpecName: "kube-api-access-lxp5l") pod "7763943d-025a-47e6-9afc-b8b02d21c876" (UID: "7763943d-025a-47e6-9afc-b8b02d21c876"). InnerVolumeSpecName "kube-api-access-lxp5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.892608 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxp5l\" (UniqueName: \"kubernetes.io/projected/7763943d-025a-47e6-9afc-b8b02d21c876-kube-api-access-lxp5l\") on node \"crc\" DevicePath \"\"" Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.892659 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.898279 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7763943d-025a-47e6-9afc-b8b02d21c876" (UID: "7763943d-025a-47e6-9afc-b8b02d21c876"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:09:42 crc kubenswrapper[4935]: I1005 10:09:42.995264 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7763943d-025a-47e6-9afc-b8b02d21c876-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.020798 4935 generic.go:334] "Generic (PLEG): container finished" podID="7763943d-025a-47e6-9afc-b8b02d21c876" containerID="476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d" exitCode=0 Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.020859 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lkx" event={"ID":"7763943d-025a-47e6-9afc-b8b02d21c876","Type":"ContainerDied","Data":"476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d"} Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.020923 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8lkx" event={"ID":"7763943d-025a-47e6-9afc-b8b02d21c876","Type":"ContainerDied","Data":"7379c3969393b7ccb437891b6d8dcff1d0975749507c2e3017e97838a933e647"} Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.020942 4935 scope.go:117] "RemoveContainer" containerID="476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.020937 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8lkx" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.066251 4935 scope.go:117] "RemoveContainer" containerID="a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.067712 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k8lkx"] Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.078101 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k8lkx"] Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.104744 4935 scope.go:117] "RemoveContainer" containerID="4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.178096 4935 scope.go:117] "RemoveContainer" containerID="476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d" Oct 05 10:09:43 crc kubenswrapper[4935]: E1005 10:09:43.178710 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d\": container with ID starting with 476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d not found: ID does not exist" containerID="476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.178764 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d"} err="failed to get container status \"476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d\": rpc error: code = NotFound desc = could not find container \"476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d\": container with ID starting with 476305737a5aeb4a8cd05121feb11cd308611bca5e53cca79f99d39ba5ff152d not found: ID does not exist" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.178796 4935 scope.go:117] "RemoveContainer" containerID="a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09" Oct 05 10:09:43 crc kubenswrapper[4935]: E1005 10:09:43.179325 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09\": container with ID starting with a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09 not found: ID does not exist" containerID="a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.179373 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09"} err="failed to get container status \"a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09\": rpc error: code = NotFound desc = could not find container \"a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09\": container with ID starting with a39fec9af0d0f010d14c787235021fdf9e86a3345b5ee175ab54cf5ed2240e09 not found: ID does not exist" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.179402 4935 scope.go:117] "RemoveContainer" containerID="4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e" Oct 05 10:09:43 crc kubenswrapper[4935]: E1005 10:09:43.179734 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e\": container with ID starting with 4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e not found: ID does not exist" containerID="4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e" Oct 05 10:09:43 crc kubenswrapper[4935]: I1005 10:09:43.179774 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e"} err="failed to get container status \"4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e\": rpc error: code = NotFound desc = could not find container \"4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e\": container with ID starting with 4c8a15972ff3a4784cf23effa70f793f6cde0795b829ef0c34b2df2f54e16f7e not found: ID does not exist" Oct 05 10:09:44 crc kubenswrapper[4935]: I1005 10:09:44.804087 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" path="/var/lib/kubelet/pods/7763943d-025a-47e6-9afc-b8b02d21c876/volumes" Oct 05 10:09:57 crc kubenswrapper[4935]: I1005 10:09:57.778534 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:09:57 crc kubenswrapper[4935]: E1005 10:09:57.779828 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:10:08 crc kubenswrapper[4935]: I1005 10:10:08.777620 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:10:08 crc kubenswrapper[4935]: E1005 10:10:08.778401 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:10:22 crc kubenswrapper[4935]: I1005 10:10:22.778564 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:10:22 crc kubenswrapper[4935]: E1005 10:10:22.779646 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:10:35 crc kubenswrapper[4935]: I1005 10:10:35.778193 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:10:35 crc kubenswrapper[4935]: E1005 10:10:35.779195 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:10:49 crc kubenswrapper[4935]: I1005 10:10:49.778949 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:10:49 crc kubenswrapper[4935]: E1005 10:10:49.779785 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:11:04 crc kubenswrapper[4935]: I1005 10:11:04.778519 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:11:04 crc kubenswrapper[4935]: E1005 10:11:04.779271 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:11:19 crc kubenswrapper[4935]: I1005 10:11:19.778856 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:11:19 crc kubenswrapper[4935]: E1005 10:11:19.780028 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:11:34 crc kubenswrapper[4935]: I1005 10:11:34.779679 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:11:34 crc kubenswrapper[4935]: E1005 10:11:34.781013 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:11:45 crc kubenswrapper[4935]: I1005 10:11:45.777125 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:11:45 crc kubenswrapper[4935]: E1005 10:11:45.778032 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:11:57 crc kubenswrapper[4935]: I1005 10:11:57.777845 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:11:57 crc kubenswrapper[4935]: E1005 10:11:57.778861 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:12:12 crc kubenswrapper[4935]: I1005 10:12:12.780968 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:12:12 crc kubenswrapper[4935]: E1005 10:12:12.782039 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.243908 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-559rv"] Oct 05 10:12:18 crc kubenswrapper[4935]: E1005 10:12:18.244813 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="extract-utilities" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.244830 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="extract-utilities" Oct 05 10:12:18 crc kubenswrapper[4935]: E1005 10:12:18.244857 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="extract-content" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.244863 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="extract-content" Oct 05 10:12:18 crc kubenswrapper[4935]: E1005 10:12:18.244923 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="registry-server" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.244932 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="registry-server" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.245136 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="7763943d-025a-47e6-9afc-b8b02d21c876" containerName="registry-server" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.246803 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.270160 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-559rv"] Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.349971 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc6k2\" (UniqueName: \"kubernetes.io/projected/dc6c4a0a-7e4f-4159-8535-297224b19d08-kube-api-access-bc6k2\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.350098 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-utilities\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.350292 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-catalog-content\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.452771 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-utilities\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.452984 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-catalog-content\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.453040 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc6k2\" (UniqueName: \"kubernetes.io/projected/dc6c4a0a-7e4f-4159-8535-297224b19d08-kube-api-access-bc6k2\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.453770 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-utilities\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.454034 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-catalog-content\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.479857 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc6k2\" (UniqueName: \"kubernetes.io/projected/dc6c4a0a-7e4f-4159-8535-297224b19d08-kube-api-access-bc6k2\") pod \"redhat-marketplace-559rv\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:18 crc kubenswrapper[4935]: I1005 10:12:18.572234 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:19 crc kubenswrapper[4935]: I1005 10:12:19.235266 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-559rv"] Oct 05 10:12:20 crc kubenswrapper[4935]: I1005 10:12:20.049919 4935 generic.go:334] "Generic (PLEG): container finished" podID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerID="f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4" exitCode=0 Oct 05 10:12:20 crc kubenswrapper[4935]: I1005 10:12:20.050381 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-559rv" event={"ID":"dc6c4a0a-7e4f-4159-8535-297224b19d08","Type":"ContainerDied","Data":"f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4"} Oct 05 10:12:20 crc kubenswrapper[4935]: I1005 10:12:20.050723 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-559rv" event={"ID":"dc6c4a0a-7e4f-4159-8535-297224b19d08","Type":"ContainerStarted","Data":"3020b3bdc96e2f6f68d8c5421ead736512b38bfb06028f9d25f83086f737fa3f"} Oct 05 10:12:20 crc kubenswrapper[4935]: I1005 10:12:20.053883 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 10:12:21 crc kubenswrapper[4935]: I1005 10:12:21.066033 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-559rv" event={"ID":"dc6c4a0a-7e4f-4159-8535-297224b19d08","Type":"ContainerStarted","Data":"d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9"} Oct 05 10:12:22 crc kubenswrapper[4935]: I1005 10:12:22.079541 4935 generic.go:334] "Generic (PLEG): container finished" podID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerID="d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9" exitCode=0 Oct 05 10:12:22 crc kubenswrapper[4935]: I1005 10:12:22.079674 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-559rv" event={"ID":"dc6c4a0a-7e4f-4159-8535-297224b19d08","Type":"ContainerDied","Data":"d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9"} Oct 05 10:12:23 crc kubenswrapper[4935]: I1005 10:12:23.095820 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-559rv" event={"ID":"dc6c4a0a-7e4f-4159-8535-297224b19d08","Type":"ContainerStarted","Data":"a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74"} Oct 05 10:12:23 crc kubenswrapper[4935]: I1005 10:12:23.132412 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-559rv" podStartSLOduration=2.724289694 podStartE2EDuration="5.132388787s" podCreationTimestamp="2025-10-05 10:12:18 +0000 UTC" firstStartedPulling="2025-10-05 10:12:20.053622146 +0000 UTC m=+11973.936248596" lastFinishedPulling="2025-10-05 10:12:22.461721229 +0000 UTC m=+11976.344347689" observedRunningTime="2025-10-05 10:12:23.12118406 +0000 UTC m=+11977.003810540" watchObservedRunningTime="2025-10-05 10:12:23.132388787 +0000 UTC m=+11977.015015247" Oct 05 10:12:26 crc kubenswrapper[4935]: I1005 10:12:26.784756 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:12:26 crc kubenswrapper[4935]: E1005 10:12:26.785605 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:12:28 crc kubenswrapper[4935]: I1005 10:12:28.572601 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:28 crc kubenswrapper[4935]: I1005 10:12:28.572980 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:28 crc kubenswrapper[4935]: I1005 10:12:28.627920 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:29 crc kubenswrapper[4935]: I1005 10:12:29.215532 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:29 crc kubenswrapper[4935]: I1005 10:12:29.298275 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-559rv"] Oct 05 10:12:31 crc kubenswrapper[4935]: I1005 10:12:31.186850 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-559rv" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerName="registry-server" containerID="cri-o://a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74" gracePeriod=2 Oct 05 10:12:31 crc kubenswrapper[4935]: I1005 10:12:31.977621 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.071754 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-utilities\") pod \"dc6c4a0a-7e4f-4159-8535-297224b19d08\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.072052 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc6k2\" (UniqueName: \"kubernetes.io/projected/dc6c4a0a-7e4f-4159-8535-297224b19d08-kube-api-access-bc6k2\") pod \"dc6c4a0a-7e4f-4159-8535-297224b19d08\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.072635 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-catalog-content\") pod \"dc6c4a0a-7e4f-4159-8535-297224b19d08\" (UID: \"dc6c4a0a-7e4f-4159-8535-297224b19d08\") " Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.075474 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-utilities" (OuterVolumeSpecName: "utilities") pod "dc6c4a0a-7e4f-4159-8535-297224b19d08" (UID: "dc6c4a0a-7e4f-4159-8535-297224b19d08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.085043 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc6c4a0a-7e4f-4159-8535-297224b19d08-kube-api-access-bc6k2" (OuterVolumeSpecName: "kube-api-access-bc6k2") pod "dc6c4a0a-7e4f-4159-8535-297224b19d08" (UID: "dc6c4a0a-7e4f-4159-8535-297224b19d08"). InnerVolumeSpecName "kube-api-access-bc6k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.091509 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc6c4a0a-7e4f-4159-8535-297224b19d08" (UID: "dc6c4a0a-7e4f-4159-8535-297224b19d08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.176353 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.176415 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc6c4a0a-7e4f-4159-8535-297224b19d08-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.176434 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc6k2\" (UniqueName: \"kubernetes.io/projected/dc6c4a0a-7e4f-4159-8535-297224b19d08-kube-api-access-bc6k2\") on node \"crc\" DevicePath \"\"" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.202861 4935 generic.go:334] "Generic (PLEG): container finished" podID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerID="a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74" exitCode=0 Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.202933 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-559rv" event={"ID":"dc6c4a0a-7e4f-4159-8535-297224b19d08","Type":"ContainerDied","Data":"a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74"} Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.202964 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-559rv" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.202998 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-559rv" event={"ID":"dc6c4a0a-7e4f-4159-8535-297224b19d08","Type":"ContainerDied","Data":"3020b3bdc96e2f6f68d8c5421ead736512b38bfb06028f9d25f83086f737fa3f"} Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.203055 4935 scope.go:117] "RemoveContainer" containerID="a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.227097 4935 scope.go:117] "RemoveContainer" containerID="d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.273567 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-559rv"] Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.289365 4935 scope.go:117] "RemoveContainer" containerID="f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.297875 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-559rv"] Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.333344 4935 scope.go:117] "RemoveContainer" containerID="a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74" Oct 05 10:12:32 crc kubenswrapper[4935]: E1005 10:12:32.334189 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74\": container with ID starting with a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74 not found: ID does not exist" containerID="a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.334259 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74"} err="failed to get container status \"a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74\": rpc error: code = NotFound desc = could not find container \"a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74\": container with ID starting with a9ff737c8d441b2bff054bd60359ead82cc563f51a9abfc42192e28c9024ec74 not found: ID does not exist" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.334298 4935 scope.go:117] "RemoveContainer" containerID="d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9" Oct 05 10:12:32 crc kubenswrapper[4935]: E1005 10:12:32.343419 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9\": container with ID starting with d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9 not found: ID does not exist" containerID="d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.343490 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9"} err="failed to get container status \"d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9\": rpc error: code = NotFound desc = could not find container \"d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9\": container with ID starting with d77bce7c2371d7b851c41c747264015f4b7ef839fcf1557fd7857785fb3b13b9 not found: ID does not exist" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.343538 4935 scope.go:117] "RemoveContainer" containerID="f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4" Oct 05 10:12:32 crc kubenswrapper[4935]: E1005 10:12:32.345655 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4\": container with ID starting with f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4 not found: ID does not exist" containerID="f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.345722 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4"} err="failed to get container status \"f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4\": rpc error: code = NotFound desc = could not find container \"f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4\": container with ID starting with f3392256f698022638cea8084bcb874c525e4a794e875717a489d1c43a4560c4 not found: ID does not exist" Oct 05 10:12:32 crc kubenswrapper[4935]: I1005 10:12:32.790751 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" path="/var/lib/kubelet/pods/dc6c4a0a-7e4f-4159-8535-297224b19d08/volumes" Oct 05 10:12:38 crc kubenswrapper[4935]: I1005 10:12:38.779216 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:12:38 crc kubenswrapper[4935]: E1005 10:12:38.780458 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:12:53 crc kubenswrapper[4935]: I1005 10:12:53.777247 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:12:53 crc kubenswrapper[4935]: E1005 10:12:53.778182 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:13:08 crc kubenswrapper[4935]: I1005 10:13:08.782466 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:13:08 crc kubenswrapper[4935]: E1005 10:13:08.783842 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:13:22 crc kubenswrapper[4935]: I1005 10:13:22.778386 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:13:22 crc kubenswrapper[4935]: E1005 10:13:22.779775 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:13:33 crc kubenswrapper[4935]: I1005 10:13:33.778387 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:13:33 crc kubenswrapper[4935]: E1005 10:13:33.779590 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:13:47 crc kubenswrapper[4935]: I1005 10:13:47.778169 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:13:47 crc kubenswrapper[4935]: E1005 10:13:47.779175 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:14:01 crc kubenswrapper[4935]: I1005 10:14:01.777187 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:14:01 crc kubenswrapper[4935]: E1005 10:14:01.778052 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:14:15 crc kubenswrapper[4935]: I1005 10:14:15.778501 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:14:16 crc kubenswrapper[4935]: I1005 10:14:16.694533 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"f78af9fbc702eb21bee411b4f79c84d0b68f0f58adccde2f1ca819901bceeda4"} Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.169768 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch"] Oct 05 10:15:00 crc kubenswrapper[4935]: E1005 10:15:00.172052 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerName="extract-utilities" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.172153 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerName="extract-utilities" Oct 05 10:15:00 crc kubenswrapper[4935]: E1005 10:15:00.172228 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerName="extract-content" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.172285 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerName="extract-content" Oct 05 10:15:00 crc kubenswrapper[4935]: E1005 10:15:00.172377 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerName="registry-server" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.172437 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerName="registry-server" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.172722 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc6c4a0a-7e4f-4159-8535-297224b19d08" containerName="registry-server" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.173584 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.179555 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.179833 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.186191 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch"] Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.240615 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swp6p\" (UniqueName: \"kubernetes.io/projected/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-kube-api-access-swp6p\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.241420 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-config-volume\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.241722 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-secret-volume\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.344349 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swp6p\" (UniqueName: \"kubernetes.io/projected/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-kube-api-access-swp6p\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.344498 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-config-volume\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.344615 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-secret-volume\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.346809 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-config-volume\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.358026 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-secret-volume\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.363328 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swp6p\" (UniqueName: \"kubernetes.io/projected/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-kube-api-access-swp6p\") pod \"collect-profiles-29327655-fc8ch\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:00 crc kubenswrapper[4935]: I1005 10:15:00.512721 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:01 crc kubenswrapper[4935]: I1005 10:15:01.040272 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch"] Oct 05 10:15:01 crc kubenswrapper[4935]: I1005 10:15:01.306673 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" event={"ID":"5de1da5c-77f1-4bdd-beb9-938fdd224f5a","Type":"ContainerStarted","Data":"65ff42b003da5e8ecf9d2639feefe7047eed9bc35824b880d8a0f32e0a847385"} Oct 05 10:15:01 crc kubenswrapper[4935]: I1005 10:15:01.307316 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" event={"ID":"5de1da5c-77f1-4bdd-beb9-938fdd224f5a","Type":"ContainerStarted","Data":"aeedaf79d2a508399a0c215c47dba0d91b3fa683643374ba57c91fb898c47f9c"} Oct 05 10:15:01 crc kubenswrapper[4935]: I1005 10:15:01.328683 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" podStartSLOduration=1.328652285 podStartE2EDuration="1.328652285s" podCreationTimestamp="2025-10-05 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 10:15:01.32621534 +0000 UTC m=+12135.208841840" watchObservedRunningTime="2025-10-05 10:15:01.328652285 +0000 UTC m=+12135.211278755" Oct 05 10:15:02 crc kubenswrapper[4935]: I1005 10:15:02.322846 4935 generic.go:334] "Generic (PLEG): container finished" podID="5de1da5c-77f1-4bdd-beb9-938fdd224f5a" containerID="65ff42b003da5e8ecf9d2639feefe7047eed9bc35824b880d8a0f32e0a847385" exitCode=0 Oct 05 10:15:02 crc kubenswrapper[4935]: I1005 10:15:02.323038 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" event={"ID":"5de1da5c-77f1-4bdd-beb9-938fdd224f5a","Type":"ContainerDied","Data":"65ff42b003da5e8ecf9d2639feefe7047eed9bc35824b880d8a0f32e0a847385"} Oct 05 10:15:03 crc kubenswrapper[4935]: I1005 10:15:03.950983 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:03 crc kubenswrapper[4935]: I1005 10:15:03.960263 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-secret-volume\") pod \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " Oct 05 10:15:03 crc kubenswrapper[4935]: I1005 10:15:03.960430 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-config-volume\") pod \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " Oct 05 10:15:03 crc kubenswrapper[4935]: I1005 10:15:03.960499 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swp6p\" (UniqueName: \"kubernetes.io/projected/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-kube-api-access-swp6p\") pod \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\" (UID: \"5de1da5c-77f1-4bdd-beb9-938fdd224f5a\") " Oct 05 10:15:03 crc kubenswrapper[4935]: I1005 10:15:03.961412 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-config-volume" (OuterVolumeSpecName: "config-volume") pod "5de1da5c-77f1-4bdd-beb9-938fdd224f5a" (UID: "5de1da5c-77f1-4bdd-beb9-938fdd224f5a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 10:15:03 crc kubenswrapper[4935]: I1005 10:15:03.970187 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5de1da5c-77f1-4bdd-beb9-938fdd224f5a" (UID: "5de1da5c-77f1-4bdd-beb9-938fdd224f5a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:15:03 crc kubenswrapper[4935]: I1005 10:15:03.971977 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-kube-api-access-swp6p" (OuterVolumeSpecName: "kube-api-access-swp6p") pod "5de1da5c-77f1-4bdd-beb9-938fdd224f5a" (UID: "5de1da5c-77f1-4bdd-beb9-938fdd224f5a"). InnerVolumeSpecName "kube-api-access-swp6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.062386 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swp6p\" (UniqueName: \"kubernetes.io/projected/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-kube-api-access-swp6p\") on node \"crc\" DevicePath \"\"" Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.063266 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.063366 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5de1da5c-77f1-4bdd-beb9-938fdd224f5a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.361363 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" event={"ID":"5de1da5c-77f1-4bdd-beb9-938fdd224f5a","Type":"ContainerDied","Data":"aeedaf79d2a508399a0c215c47dba0d91b3fa683643374ba57c91fb898c47f9c"} Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.361408 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aeedaf79d2a508399a0c215c47dba0d91b3fa683643374ba57c91fb898c47f9c" Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.361471 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327655-fc8ch" Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.431105 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s"] Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.440098 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327610-xnl6s"] Oct 05 10:15:04 crc kubenswrapper[4935]: I1005 10:15:04.793834 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d654093-f8b9-4333-8949-52bd35aea8d1" path="/var/lib/kubelet/pods/8d654093-f8b9-4333-8949-52bd35aea8d1/volumes" Oct 05 10:15:05 crc kubenswrapper[4935]: I1005 10:15:05.982116 4935 scope.go:117] "RemoveContainer" containerID="cd27f460da8ed74c0ffd2dfc66aa734a65fa5751fd1a5f03a675f8f8fcbc59d6" Oct 05 10:16:44 crc kubenswrapper[4935]: I1005 10:16:44.289123 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:16:44 crc kubenswrapper[4935]: I1005 10:16:44.289924 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:17:00 crc kubenswrapper[4935]: I1005 10:17:00.881719 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vvdfp"] Oct 05 10:17:00 crc kubenswrapper[4935]: E1005 10:17:00.882767 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de1da5c-77f1-4bdd-beb9-938fdd224f5a" containerName="collect-profiles" Oct 05 10:17:00 crc kubenswrapper[4935]: I1005 10:17:00.882785 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de1da5c-77f1-4bdd-beb9-938fdd224f5a" containerName="collect-profiles" Oct 05 10:17:00 crc kubenswrapper[4935]: I1005 10:17:00.883098 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de1da5c-77f1-4bdd-beb9-938fdd224f5a" containerName="collect-profiles" Oct 05 10:17:00 crc kubenswrapper[4935]: I1005 10:17:00.885143 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:00 crc kubenswrapper[4935]: I1005 10:17:00.902028 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vvdfp"] Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.015810 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2vnc\" (UniqueName: \"kubernetes.io/projected/8fda3e4a-a030-457d-aabc-af8888e57c8e-kube-api-access-v2vnc\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.016316 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-utilities\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.016485 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-catalog-content\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.119178 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-utilities\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.119500 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-catalog-content\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.119922 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2vnc\" (UniqueName: \"kubernetes.io/projected/8fda3e4a-a030-457d-aabc-af8888e57c8e-kube-api-access-v2vnc\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.119994 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-utilities\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.120178 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-catalog-content\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.144684 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2vnc\" (UniqueName: \"kubernetes.io/projected/8fda3e4a-a030-457d-aabc-af8888e57c8e-kube-api-access-v2vnc\") pod \"community-operators-vvdfp\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.234148 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:01 crc kubenswrapper[4935]: I1005 10:17:01.900702 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vvdfp"] Oct 05 10:17:02 crc kubenswrapper[4935]: I1005 10:17:02.881689 4935 generic.go:334] "Generic (PLEG): container finished" podID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerID="ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310" exitCode=0 Oct 05 10:17:02 crc kubenswrapper[4935]: I1005 10:17:02.881762 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvdfp" event={"ID":"8fda3e4a-a030-457d-aabc-af8888e57c8e","Type":"ContainerDied","Data":"ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310"} Oct 05 10:17:02 crc kubenswrapper[4935]: I1005 10:17:02.882081 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvdfp" event={"ID":"8fda3e4a-a030-457d-aabc-af8888e57c8e","Type":"ContainerStarted","Data":"85e1593c0ae895e894e989c69aa8ba0a50fe7c086fa824e0dd9a41cf3836e237"} Oct 05 10:17:03 crc kubenswrapper[4935]: I1005 10:17:03.897070 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvdfp" event={"ID":"8fda3e4a-a030-457d-aabc-af8888e57c8e","Type":"ContainerStarted","Data":"2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c"} Oct 05 10:17:05 crc kubenswrapper[4935]: I1005 10:17:05.928207 4935 generic.go:334] "Generic (PLEG): container finished" podID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerID="2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c" exitCode=0 Oct 05 10:17:05 crc kubenswrapper[4935]: I1005 10:17:05.928489 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvdfp" event={"ID":"8fda3e4a-a030-457d-aabc-af8888e57c8e","Type":"ContainerDied","Data":"2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c"} Oct 05 10:17:06 crc kubenswrapper[4935]: I1005 10:17:06.954860 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvdfp" event={"ID":"8fda3e4a-a030-457d-aabc-af8888e57c8e","Type":"ContainerStarted","Data":"476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501"} Oct 05 10:17:06 crc kubenswrapper[4935]: I1005 10:17:06.979630 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vvdfp" podStartSLOduration=3.445534338 podStartE2EDuration="6.979608216s" podCreationTimestamp="2025-10-05 10:17:00 +0000 UTC" firstStartedPulling="2025-10-05 10:17:02.886387912 +0000 UTC m=+12256.769014402" lastFinishedPulling="2025-10-05 10:17:06.42046182 +0000 UTC m=+12260.303088280" observedRunningTime="2025-10-05 10:17:06.972698273 +0000 UTC m=+12260.855324743" watchObservedRunningTime="2025-10-05 10:17:06.979608216 +0000 UTC m=+12260.862234676" Oct 05 10:17:11 crc kubenswrapper[4935]: I1005 10:17:11.236160 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:11 crc kubenswrapper[4935]: I1005 10:17:11.237068 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:11 crc kubenswrapper[4935]: I1005 10:17:11.326227 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:12 crc kubenswrapper[4935]: I1005 10:17:12.095217 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:12 crc kubenswrapper[4935]: I1005 10:17:12.163013 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vvdfp"] Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.052498 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vvdfp" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerName="registry-server" containerID="cri-o://476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501" gracePeriod=2 Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.291593 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.291727 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.630324 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.769613 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2vnc\" (UniqueName: \"kubernetes.io/projected/8fda3e4a-a030-457d-aabc-af8888e57c8e-kube-api-access-v2vnc\") pod \"8fda3e4a-a030-457d-aabc-af8888e57c8e\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.769885 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-catalog-content\") pod \"8fda3e4a-a030-457d-aabc-af8888e57c8e\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.769967 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-utilities\") pod \"8fda3e4a-a030-457d-aabc-af8888e57c8e\" (UID: \"8fda3e4a-a030-457d-aabc-af8888e57c8e\") " Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.771878 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-utilities" (OuterVolumeSpecName: "utilities") pod "8fda3e4a-a030-457d-aabc-af8888e57c8e" (UID: "8fda3e4a-a030-457d-aabc-af8888e57c8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.776399 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fda3e4a-a030-457d-aabc-af8888e57c8e-kube-api-access-v2vnc" (OuterVolumeSpecName: "kube-api-access-v2vnc") pod "8fda3e4a-a030-457d-aabc-af8888e57c8e" (UID: "8fda3e4a-a030-457d-aabc-af8888e57c8e"). InnerVolumeSpecName "kube-api-access-v2vnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.831832 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fda3e4a-a030-457d-aabc-af8888e57c8e" (UID: "8fda3e4a-a030-457d-aabc-af8888e57c8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.872886 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2vnc\" (UniqueName: \"kubernetes.io/projected/8fda3e4a-a030-457d-aabc-af8888e57c8e-kube-api-access-v2vnc\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.872964 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:14 crc kubenswrapper[4935]: I1005 10:17:14.872984 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fda3e4a-a030-457d-aabc-af8888e57c8e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.069566 4935 generic.go:334] "Generic (PLEG): container finished" podID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerID="476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501" exitCode=0 Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.069742 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvdfp" event={"ID":"8fda3e4a-a030-457d-aabc-af8888e57c8e","Type":"ContainerDied","Data":"476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501"} Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.071104 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvdfp" event={"ID":"8fda3e4a-a030-457d-aabc-af8888e57c8e","Type":"ContainerDied","Data":"85e1593c0ae895e894e989c69aa8ba0a50fe7c086fa824e0dd9a41cf3836e237"} Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.071159 4935 scope.go:117] "RemoveContainer" containerID="476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.069851 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvdfp" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.127559 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vvdfp"] Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.130925 4935 scope.go:117] "RemoveContainer" containerID="2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.137741 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vvdfp"] Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.172314 4935 scope.go:117] "RemoveContainer" containerID="ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.231164 4935 scope.go:117] "RemoveContainer" containerID="476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501" Oct 05 10:17:15 crc kubenswrapper[4935]: E1005 10:17:15.241928 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501\": container with ID starting with 476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501 not found: ID does not exist" containerID="476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.242037 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501"} err="failed to get container status \"476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501\": rpc error: code = NotFound desc = could not find container \"476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501\": container with ID starting with 476fa583611b51c5e401e6fc8fcdd0a4e0f2ba6949c0fba3179ffc724695c501 not found: ID does not exist" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.242091 4935 scope.go:117] "RemoveContainer" containerID="2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c" Oct 05 10:17:15 crc kubenswrapper[4935]: E1005 10:17:15.242801 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c\": container with ID starting with 2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c not found: ID does not exist" containerID="2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.242863 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c"} err="failed to get container status \"2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c\": rpc error: code = NotFound desc = could not find container \"2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c\": container with ID starting with 2b31c204f085e5c719ebe52c4c73757a391b4d10ce1c069b3462d28cfad6b89c not found: ID does not exist" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.242909 4935 scope.go:117] "RemoveContainer" containerID="ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310" Oct 05 10:17:15 crc kubenswrapper[4935]: E1005 10:17:15.243179 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310\": container with ID starting with ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310 not found: ID does not exist" containerID="ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310" Oct 05 10:17:15 crc kubenswrapper[4935]: I1005 10:17:15.243199 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310"} err="failed to get container status \"ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310\": rpc error: code = NotFound desc = could not find container \"ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310\": container with ID starting with ae89d78c2d407bd94383e32749067eeadbcccc6a583925b29d4f5e3db1b0b310 not found: ID does not exist" Oct 05 10:17:16 crc kubenswrapper[4935]: I1005 10:17:16.798658 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" path="/var/lib/kubelet/pods/8fda3e4a-a030-457d-aabc-af8888e57c8e/volumes" Oct 05 10:17:31 crc kubenswrapper[4935]: I1005 10:17:31.292789 4935 generic.go:334] "Generic (PLEG): container finished" podID="c54822d6-7348-4d81-8681-f338af0badc1" containerID="75c1fa9680d767c40d2f9685ede1d068a8a5e0f33fad8d26fb898946eacedf54" exitCode=1 Oct 05 10:17:31 crc kubenswrapper[4935]: I1005 10:17:31.292856 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c54822d6-7348-4d81-8681-f338af0badc1","Type":"ContainerDied","Data":"75c1fa9680d767c40d2f9685ede1d068a8a5e0f33fad8d26fb898946eacedf54"} Oct 05 10:17:32 crc kubenswrapper[4935]: I1005 10:17:32.867560 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.057367 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.057962 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ssh-key\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.058055 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-temporary\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.058172 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt7r7\" (UniqueName: \"kubernetes.io/projected/c54822d6-7348-4d81-8681-f338af0badc1-kube-api-access-pt7r7\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.058241 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ca-certs\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.058310 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-workdir\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.058374 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-config-data\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.058429 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config-secret\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.058524 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"c54822d6-7348-4d81-8681-f338af0badc1\" (UID: \"c54822d6-7348-4d81-8681-f338af0badc1\") " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.060272 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.063050 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-config-data" (OuterVolumeSpecName: "config-data") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.067744 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.071114 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c54822d6-7348-4d81-8681-f338af0badc1-kube-api-access-pt7r7" (OuterVolumeSpecName: "kube-api-access-pt7r7") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "kube-api-access-pt7r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.072610 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.114696 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.115323 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.118781 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.159479 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c54822d6-7348-4d81-8681-f338af0badc1" (UID: "c54822d6-7348-4d81-8681-f338af0badc1"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162659 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162707 4935 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162727 4935 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162748 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt7r7\" (UniqueName: \"kubernetes.io/projected/c54822d6-7348-4d81-8681-f338af0badc1-kube-api-access-pt7r7\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162767 4935 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162790 4935 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c54822d6-7348-4d81-8681-f338af0badc1-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162807 4935 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c54822d6-7348-4d81-8681-f338af0badc1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162824 4935 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54822d6-7348-4d81-8681-f338af0badc1-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.162865 4935 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.208695 4935 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.265511 4935 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.320014 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c54822d6-7348-4d81-8681-f338af0badc1","Type":"ContainerDied","Data":"c75ffa72d41ef6b986ee5c85f13c264e919cffbccbd465537b2c6e86d1e7714a"} Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.320067 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 10:17:33 crc kubenswrapper[4935]: I1005 10:17:33.320092 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c75ffa72d41ef6b986ee5c85f13c264e919cffbccbd465537b2c6e86d1e7714a" Oct 05 10:17:44 crc kubenswrapper[4935]: I1005 10:17:44.290461 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:17:44 crc kubenswrapper[4935]: I1005 10:17:44.291379 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:17:44 crc kubenswrapper[4935]: I1005 10:17:44.291469 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 10:17:44 crc kubenswrapper[4935]: I1005 10:17:44.292808 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f78af9fbc702eb21bee411b4f79c84d0b68f0f58adccde2f1ca819901bceeda4"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 10:17:44 crc kubenswrapper[4935]: I1005 10:17:44.292941 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://f78af9fbc702eb21bee411b4f79c84d0b68f0f58adccde2f1ca819901bceeda4" gracePeriod=600 Oct 05 10:17:44 crc kubenswrapper[4935]: I1005 10:17:44.476290 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="f78af9fbc702eb21bee411b4f79c84d0b68f0f58adccde2f1ca819901bceeda4" exitCode=0 Oct 05 10:17:44 crc kubenswrapper[4935]: I1005 10:17:44.476361 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"f78af9fbc702eb21bee411b4f79c84d0b68f0f58adccde2f1ca819901bceeda4"} Oct 05 10:17:44 crc kubenswrapper[4935]: I1005 10:17:44.476516 4935 scope.go:117] "RemoveContainer" containerID="e27c9bf194b582d92e6ffe2717a41b22f450dc8e4055f2fe4200306834cc47b4" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.348596 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 10:17:45 crc kubenswrapper[4935]: E1005 10:17:45.349956 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54822d6-7348-4d81-8681-f338af0badc1" containerName="tempest-tests-tempest-tests-runner" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.349979 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54822d6-7348-4d81-8681-f338af0badc1" containerName="tempest-tests-tempest-tests-runner" Oct 05 10:17:45 crc kubenswrapper[4935]: E1005 10:17:45.350007 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerName="extract-content" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.350018 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerName="extract-content" Oct 05 10:17:45 crc kubenswrapper[4935]: E1005 10:17:45.350054 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerName="extract-utilities" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.350065 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerName="extract-utilities" Oct 05 10:17:45 crc kubenswrapper[4935]: E1005 10:17:45.350123 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerName="registry-server" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.350134 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerName="registry-server" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.350498 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fda3e4a-a030-457d-aabc-af8888e57c8e" containerName="registry-server" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.350531 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54822d6-7348-4d81-8681-f338af0badc1" containerName="tempest-tests-tempest-tests-runner" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.351688 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.356586 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sw94x" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.378122 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.454121 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0146f93c-5183-4802-aba2-0780ca2411c2\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.454535 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9gxg\" (UniqueName: \"kubernetes.io/projected/0146f93c-5183-4802-aba2-0780ca2411c2-kube-api-access-f9gxg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0146f93c-5183-4802-aba2-0780ca2411c2\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.492635 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24"} Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.557979 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0146f93c-5183-4802-aba2-0780ca2411c2\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.558091 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9gxg\" (UniqueName: \"kubernetes.io/projected/0146f93c-5183-4802-aba2-0780ca2411c2-kube-api-access-f9gxg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0146f93c-5183-4802-aba2-0780ca2411c2\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.558691 4935 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0146f93c-5183-4802-aba2-0780ca2411c2\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.582493 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9gxg\" (UniqueName: \"kubernetes.io/projected/0146f93c-5183-4802-aba2-0780ca2411c2-kube-api-access-f9gxg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0146f93c-5183-4802-aba2-0780ca2411c2\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.587159 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0146f93c-5183-4802-aba2-0780ca2411c2\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:45 crc kubenswrapper[4935]: I1005 10:17:45.673008 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 10:17:46 crc kubenswrapper[4935]: I1005 10:17:46.174651 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 10:17:46 crc kubenswrapper[4935]: I1005 10:17:46.176277 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 10:17:46 crc kubenswrapper[4935]: I1005 10:17:46.508861 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"0146f93c-5183-4802-aba2-0780ca2411c2","Type":"ContainerStarted","Data":"9e5736af2ac2cf3e7ce82546ad75aa32651d1e48a53e75094753824b506464bf"} Oct 05 10:17:47 crc kubenswrapper[4935]: I1005 10:17:47.520676 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"0146f93c-5183-4802-aba2-0780ca2411c2","Type":"ContainerStarted","Data":"abd19d3db3b46c484a053d384cf56a761221185be5e0e096078c9885c9a35747"} Oct 05 10:17:47 crc kubenswrapper[4935]: I1005 10:17:47.539009 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.481394146 podStartE2EDuration="2.53899142s" podCreationTimestamp="2025-10-05 10:17:45 +0000 UTC" firstStartedPulling="2025-10-05 10:17:46.176062771 +0000 UTC m=+12300.058689231" lastFinishedPulling="2025-10-05 10:17:47.233660015 +0000 UTC m=+12301.116286505" observedRunningTime="2025-10-05 10:17:47.53634756 +0000 UTC m=+12301.418974020" watchObservedRunningTime="2025-10-05 10:17:47.53899142 +0000 UTC m=+12301.421617880" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.338115 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9m5xl/must-gather-btd99"] Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.343309 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.345870 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-9m5xl"/"default-dockercfg-gpkvk" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.346366 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9m5xl"/"kube-root-ca.crt" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.346805 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9m5xl"/"openshift-service-ca.crt" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.366568 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9m5xl/must-gather-btd99"] Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.539053 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/94e6a683-891d-4846-987a-42fc12fe8fb5-must-gather-output\") pod \"must-gather-btd99\" (UID: \"94e6a683-891d-4846-987a-42fc12fe8fb5\") " pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.539531 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2flq6\" (UniqueName: \"kubernetes.io/projected/94e6a683-891d-4846-987a-42fc12fe8fb5-kube-api-access-2flq6\") pod \"must-gather-btd99\" (UID: \"94e6a683-891d-4846-987a-42fc12fe8fb5\") " pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.642594 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2flq6\" (UniqueName: \"kubernetes.io/projected/94e6a683-891d-4846-987a-42fc12fe8fb5-kube-api-access-2flq6\") pod \"must-gather-btd99\" (UID: \"94e6a683-891d-4846-987a-42fc12fe8fb5\") " pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.642882 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/94e6a683-891d-4846-987a-42fc12fe8fb5-must-gather-output\") pod \"must-gather-btd99\" (UID: \"94e6a683-891d-4846-987a-42fc12fe8fb5\") " pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.643321 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/94e6a683-891d-4846-987a-42fc12fe8fb5-must-gather-output\") pod \"must-gather-btd99\" (UID: \"94e6a683-891d-4846-987a-42fc12fe8fb5\") " pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.665469 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2flq6\" (UniqueName: \"kubernetes.io/projected/94e6a683-891d-4846-987a-42fc12fe8fb5-kube-api-access-2flq6\") pod \"must-gather-btd99\" (UID: \"94e6a683-891d-4846-987a-42fc12fe8fb5\") " pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:19:02 crc kubenswrapper[4935]: I1005 10:19:02.962612 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:19:03 crc kubenswrapper[4935]: I1005 10:19:03.602280 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9m5xl/must-gather-btd99"] Oct 05 10:19:04 crc kubenswrapper[4935]: I1005 10:19:04.575259 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/must-gather-btd99" event={"ID":"94e6a683-891d-4846-987a-42fc12fe8fb5","Type":"ContainerStarted","Data":"6485e9acd78ce024ae7620c9752969400ba108a1843a58f8fd00698e5df6b6fa"} Oct 05 10:19:08 crc kubenswrapper[4935]: I1005 10:19:08.627414 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/must-gather-btd99" event={"ID":"94e6a683-891d-4846-987a-42fc12fe8fb5","Type":"ContainerStarted","Data":"e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77"} Oct 05 10:19:09 crc kubenswrapper[4935]: I1005 10:19:09.650303 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/must-gather-btd99" event={"ID":"94e6a683-891d-4846-987a-42fc12fe8fb5","Type":"ContainerStarted","Data":"42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c"} Oct 05 10:19:09 crc kubenswrapper[4935]: I1005 10:19:09.677106 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9m5xl/must-gather-btd99" podStartSLOduration=2.964608136 podStartE2EDuration="7.677063285s" podCreationTimestamp="2025-10-05 10:19:02 +0000 UTC" firstStartedPulling="2025-10-05 10:19:03.613144011 +0000 UTC m=+12377.495770471" lastFinishedPulling="2025-10-05 10:19:08.32559916 +0000 UTC m=+12382.208225620" observedRunningTime="2025-10-05 10:19:09.665262372 +0000 UTC m=+12383.547888872" watchObservedRunningTime="2025-10-05 10:19:09.677063285 +0000 UTC m=+12383.559689815" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.285495 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-4s45c"] Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.288013 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.385147 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-host\") pod \"crc-debug-4s45c\" (UID: \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\") " pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.385260 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdq4h\" (UniqueName: \"kubernetes.io/projected/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-kube-api-access-hdq4h\") pod \"crc-debug-4s45c\" (UID: \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\") " pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.487330 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-host\") pod \"crc-debug-4s45c\" (UID: \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\") " pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.487404 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdq4h\" (UniqueName: \"kubernetes.io/projected/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-kube-api-access-hdq4h\") pod \"crc-debug-4s45c\" (UID: \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\") " pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.487515 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-host\") pod \"crc-debug-4s45c\" (UID: \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\") " pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.510670 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdq4h\" (UniqueName: \"kubernetes.io/projected/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-kube-api-access-hdq4h\") pod \"crc-debug-4s45c\" (UID: \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\") " pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.616977 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:19:15 crc kubenswrapper[4935]: I1005 10:19:15.742248 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/crc-debug-4s45c" event={"ID":"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2","Type":"ContainerStarted","Data":"493a49eab43668ce9e800649294e6a10b2bb661e603ef9657335b80a821509bd"} Oct 05 10:19:24 crc kubenswrapper[4935]: I1005 10:19:24.873993 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/crc-debug-4s45c" event={"ID":"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2","Type":"ContainerStarted","Data":"283de8d867f25e9e5ef8888a02dfeda250f160601db79293ff3ec803c47b4441"} Oct 05 10:19:24 crc kubenswrapper[4935]: I1005 10:19:24.905535 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9m5xl/crc-debug-4s45c" podStartSLOduration=1.124833971 podStartE2EDuration="9.905505241s" podCreationTimestamp="2025-10-05 10:19:15 +0000 UTC" firstStartedPulling="2025-10-05 10:19:15.66138763 +0000 UTC m=+12389.544014090" lastFinishedPulling="2025-10-05 10:19:24.44205886 +0000 UTC m=+12398.324685360" observedRunningTime="2025-10-05 10:19:24.888512301 +0000 UTC m=+12398.771138801" watchObservedRunningTime="2025-10-05 10:19:24.905505241 +0000 UTC m=+12398.788131711" Oct 05 10:19:44 crc kubenswrapper[4935]: I1005 10:19:44.290137 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:19:44 crc kubenswrapper[4935]: I1005 10:19:44.291447 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:20:14 crc kubenswrapper[4935]: I1005 10:20:14.289295 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:20:14 crc kubenswrapper[4935]: I1005 10:20:14.290135 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.289042 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.289847 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.289921 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.290876 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.290958 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" gracePeriod=600 Oct 05 10:20:44 crc kubenswrapper[4935]: E1005 10:20:44.424566 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.921372 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" exitCode=0 Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.921448 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24"} Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.922267 4935 scope.go:117] "RemoveContainer" containerID="f78af9fbc702eb21bee411b4f79c84d0b68f0f58adccde2f1ca819901bceeda4" Oct 05 10:20:44 crc kubenswrapper[4935]: I1005 10:20:44.923503 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:20:44 crc kubenswrapper[4935]: E1005 10:20:44.924308 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:20:56 crc kubenswrapper[4935]: I1005 10:20:56.784070 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:20:56 crc kubenswrapper[4935]: E1005 10:20:56.784759 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:21:10 crc kubenswrapper[4935]: I1005 10:21:10.777145 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:21:10 crc kubenswrapper[4935]: E1005 10:21:10.777832 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.405869 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zx7q4"] Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.410832 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.436627 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zx7q4"] Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.463344 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-utilities\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.463410 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t954\" (UniqueName: \"kubernetes.io/projected/56e353ad-7d3d-4500-8f35-af7eb281caab-kube-api-access-2t954\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.463727 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-catalog-content\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.566101 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-catalog-content\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.566236 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-utilities\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.566293 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t954\" (UniqueName: \"kubernetes.io/projected/56e353ad-7d3d-4500-8f35-af7eb281caab-kube-api-access-2t954\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.566604 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-catalog-content\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.566743 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-utilities\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.594009 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t954\" (UniqueName: \"kubernetes.io/projected/56e353ad-7d3d-4500-8f35-af7eb281caab-kube-api-access-2t954\") pod \"certified-operators-zx7q4\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:12 crc kubenswrapper[4935]: I1005 10:21:12.730205 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:13 crc kubenswrapper[4935]: I1005 10:21:13.283877 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zx7q4"] Oct 05 10:21:13 crc kubenswrapper[4935]: I1005 10:21:13.340065 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx7q4" event={"ID":"56e353ad-7d3d-4500-8f35-af7eb281caab","Type":"ContainerStarted","Data":"51e87eb0c32bfc1b8ebb4b7af4eb393d61207ef48f0f7ee192807bbb97d390ae"} Oct 05 10:21:14 crc kubenswrapper[4935]: I1005 10:21:14.354454 4935 generic.go:334] "Generic (PLEG): container finished" podID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerID="aeda3bf0f93f20beabaa106abe0bdaffc1cd1af68630c38aeefd792ce4dd2877" exitCode=0 Oct 05 10:21:14 crc kubenswrapper[4935]: I1005 10:21:14.354694 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx7q4" event={"ID":"56e353ad-7d3d-4500-8f35-af7eb281caab","Type":"ContainerDied","Data":"aeda3bf0f93f20beabaa106abe0bdaffc1cd1af68630c38aeefd792ce4dd2877"} Oct 05 10:21:15 crc kubenswrapper[4935]: I1005 10:21:15.366317 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx7q4" event={"ID":"56e353ad-7d3d-4500-8f35-af7eb281caab","Type":"ContainerStarted","Data":"4494c8da21a12ab9c64873483f6cb507843813664ea10f7b5a3560f1027d5763"} Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.404835 4935 generic.go:334] "Generic (PLEG): container finished" podID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerID="4494c8da21a12ab9c64873483f6cb507843813664ea10f7b5a3560f1027d5763" exitCode=0 Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.404941 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx7q4" event={"ID":"56e353ad-7d3d-4500-8f35-af7eb281caab","Type":"ContainerDied","Data":"4494c8da21a12ab9c64873483f6cb507843813664ea10f7b5a3560f1027d5763"} Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.597740 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zv6sr"] Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.599851 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.619717 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zv6sr"] Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.773366 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n298g\" (UniqueName: \"kubernetes.io/projected/d2a84e74-15df-4e59-be99-0232bf5ffaa7-kube-api-access-n298g\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.773654 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-utilities\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.774203 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-catalog-content\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.876429 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-catalog-content\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.876529 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n298g\" (UniqueName: \"kubernetes.io/projected/d2a84e74-15df-4e59-be99-0232bf5ffaa7-kube-api-access-n298g\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.876630 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-utilities\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.877099 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-catalog-content\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.877135 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-utilities\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.906008 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n298g\" (UniqueName: \"kubernetes.io/projected/d2a84e74-15df-4e59-be99-0232bf5ffaa7-kube-api-access-n298g\") pod \"redhat-operators-zv6sr\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:17 crc kubenswrapper[4935]: I1005 10:21:17.922377 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:18 crc kubenswrapper[4935]: I1005 10:21:18.423878 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx7q4" event={"ID":"56e353ad-7d3d-4500-8f35-af7eb281caab","Type":"ContainerStarted","Data":"81cc446f77e5b38d697aa9bcb6295d66e0688658f746c2443e1b49f8836cded1"} Oct 05 10:21:18 crc kubenswrapper[4935]: I1005 10:21:18.479512 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zv6sr"] Oct 05 10:21:18 crc kubenswrapper[4935]: I1005 10:21:18.492524 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zx7q4" podStartSLOduration=2.902519116 podStartE2EDuration="6.492503514s" podCreationTimestamp="2025-10-05 10:21:12 +0000 UTC" firstStartedPulling="2025-10-05 10:21:14.356365095 +0000 UTC m=+12508.238991555" lastFinishedPulling="2025-10-05 10:21:17.946349493 +0000 UTC m=+12511.828975953" observedRunningTime="2025-10-05 10:21:18.446478255 +0000 UTC m=+12512.329104725" watchObservedRunningTime="2025-10-05 10:21:18.492503514 +0000 UTC m=+12512.375129964" Oct 05 10:21:19 crc kubenswrapper[4935]: I1005 10:21:19.473281 4935 generic.go:334] "Generic (PLEG): container finished" podID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerID="862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7" exitCode=0 Oct 05 10:21:19 crc kubenswrapper[4935]: I1005 10:21:19.473417 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv6sr" event={"ID":"d2a84e74-15df-4e59-be99-0232bf5ffaa7","Type":"ContainerDied","Data":"862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7"} Oct 05 10:21:19 crc kubenswrapper[4935]: I1005 10:21:19.473646 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv6sr" event={"ID":"d2a84e74-15df-4e59-be99-0232bf5ffaa7","Type":"ContainerStarted","Data":"1c3fc9b8f8d15fd5168316ff58261072bd783c12baebcb4b6958ae273b7f7cb0"} Oct 05 10:21:21 crc kubenswrapper[4935]: I1005 10:21:21.498214 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv6sr" event={"ID":"d2a84e74-15df-4e59-be99-0232bf5ffaa7","Type":"ContainerStarted","Data":"6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154"} Oct 05 10:21:22 crc kubenswrapper[4935]: I1005 10:21:22.730456 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:22 crc kubenswrapper[4935]: I1005 10:21:22.730526 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:22 crc kubenswrapper[4935]: I1005 10:21:22.777978 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:21:22 crc kubenswrapper[4935]: E1005 10:21:22.778259 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:21:22 crc kubenswrapper[4935]: I1005 10:21:22.801580 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:23 crc kubenswrapper[4935]: I1005 10:21:23.600118 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:24 crc kubenswrapper[4935]: I1005 10:21:24.530476 4935 generic.go:334] "Generic (PLEG): container finished" podID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerID="6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154" exitCode=0 Oct 05 10:21:24 crc kubenswrapper[4935]: I1005 10:21:24.530547 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv6sr" event={"ID":"d2a84e74-15df-4e59-be99-0232bf5ffaa7","Type":"ContainerDied","Data":"6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154"} Oct 05 10:21:25 crc kubenswrapper[4935]: I1005 10:21:25.543659 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv6sr" event={"ID":"d2a84e74-15df-4e59-be99-0232bf5ffaa7","Type":"ContainerStarted","Data":"cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34"} Oct 05 10:21:25 crc kubenswrapper[4935]: I1005 10:21:25.561712 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zv6sr" podStartSLOduration=3.091282827 podStartE2EDuration="8.561688475s" podCreationTimestamp="2025-10-05 10:21:17 +0000 UTC" firstStartedPulling="2025-10-05 10:21:19.479055236 +0000 UTC m=+12513.361681726" lastFinishedPulling="2025-10-05 10:21:24.949460914 +0000 UTC m=+12518.832087374" observedRunningTime="2025-10-05 10:21:25.560756451 +0000 UTC m=+12519.443382911" watchObservedRunningTime="2025-10-05 10:21:25.561688475 +0000 UTC m=+12519.444314945" Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.189200 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zx7q4"] Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.189750 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zx7q4" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerName="registry-server" containerID="cri-o://81cc446f77e5b38d697aa9bcb6295d66e0688658f746c2443e1b49f8836cded1" gracePeriod=2 Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.565405 4935 generic.go:334] "Generic (PLEG): container finished" podID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerID="81cc446f77e5b38d697aa9bcb6295d66e0688658f746c2443e1b49f8836cded1" exitCode=0 Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.565734 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx7q4" event={"ID":"56e353ad-7d3d-4500-8f35-af7eb281caab","Type":"ContainerDied","Data":"81cc446f77e5b38d697aa9bcb6295d66e0688658f746c2443e1b49f8836cded1"} Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.810231 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.865691 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t954\" (UniqueName: \"kubernetes.io/projected/56e353ad-7d3d-4500-8f35-af7eb281caab-kube-api-access-2t954\") pod \"56e353ad-7d3d-4500-8f35-af7eb281caab\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.865789 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-catalog-content\") pod \"56e353ad-7d3d-4500-8f35-af7eb281caab\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.865834 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-utilities\") pod \"56e353ad-7d3d-4500-8f35-af7eb281caab\" (UID: \"56e353ad-7d3d-4500-8f35-af7eb281caab\") " Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.867658 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-utilities" (OuterVolumeSpecName: "utilities") pod "56e353ad-7d3d-4500-8f35-af7eb281caab" (UID: "56e353ad-7d3d-4500-8f35-af7eb281caab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.903259 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56e353ad-7d3d-4500-8f35-af7eb281caab-kube-api-access-2t954" (OuterVolumeSpecName: "kube-api-access-2t954") pod "56e353ad-7d3d-4500-8f35-af7eb281caab" (UID: "56e353ad-7d3d-4500-8f35-af7eb281caab"). InnerVolumeSpecName "kube-api-access-2t954". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.949965 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56e353ad-7d3d-4500-8f35-af7eb281caab" (UID: "56e353ad-7d3d-4500-8f35-af7eb281caab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.968567 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t954\" (UniqueName: \"kubernetes.io/projected/56e353ad-7d3d-4500-8f35-af7eb281caab-kube-api-access-2t954\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.968610 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:26 crc kubenswrapper[4935]: I1005 10:21:26.968623 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56e353ad-7d3d-4500-8f35-af7eb281caab-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.580311 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx7q4" event={"ID":"56e353ad-7d3d-4500-8f35-af7eb281caab","Type":"ContainerDied","Data":"51e87eb0c32bfc1b8ebb4b7af4eb393d61207ef48f0f7ee192807bbb97d390ae"} Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.580792 4935 scope.go:117] "RemoveContainer" containerID="81cc446f77e5b38d697aa9bcb6295d66e0688658f746c2443e1b49f8836cded1" Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.580404 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zx7q4" Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.607008 4935 scope.go:117] "RemoveContainer" containerID="4494c8da21a12ab9c64873483f6cb507843813664ea10f7b5a3560f1027d5763" Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.628012 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zx7q4"] Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.638486 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zx7q4"] Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.642410 4935 scope.go:117] "RemoveContainer" containerID="aeda3bf0f93f20beabaa106abe0bdaffc1cd1af68630c38aeefd792ce4dd2877" Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.923280 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:27 crc kubenswrapper[4935]: I1005 10:21:27.923339 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:28 crc kubenswrapper[4935]: I1005 10:21:28.793625 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" path="/var/lib/kubelet/pods/56e353ad-7d3d-4500-8f35-af7eb281caab/volumes" Oct 05 10:21:29 crc kubenswrapper[4935]: I1005 10:21:29.002327 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zv6sr" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="registry-server" probeResult="failure" output=< Oct 05 10:21:29 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 10:21:29 crc kubenswrapper[4935]: > Oct 05 10:21:37 crc kubenswrapper[4935]: I1005 10:21:37.778165 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:21:37 crc kubenswrapper[4935]: E1005 10:21:37.780482 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:21:38 crc kubenswrapper[4935]: I1005 10:21:38.017034 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:38 crc kubenswrapper[4935]: I1005 10:21:38.082357 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:38 crc kubenswrapper[4935]: I1005 10:21:38.270034 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zv6sr"] Oct 05 10:21:38 crc kubenswrapper[4935]: I1005 10:21:38.753024 4935 generic.go:334] "Generic (PLEG): container finished" podID="a76fb95e-5ab9-4dfb-b055-d3e0987b86a2" containerID="283de8d867f25e9e5ef8888a02dfeda250f160601db79293ff3ec803c47b4441" exitCode=0 Oct 05 10:21:38 crc kubenswrapper[4935]: I1005 10:21:38.753547 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/crc-debug-4s45c" event={"ID":"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2","Type":"ContainerDied","Data":"283de8d867f25e9e5ef8888a02dfeda250f160601db79293ff3ec803c47b4441"} Oct 05 10:21:39 crc kubenswrapper[4935]: I1005 10:21:39.765960 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zv6sr" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="registry-server" containerID="cri-o://cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34" gracePeriod=2 Oct 05 10:21:39 crc kubenswrapper[4935]: I1005 10:21:39.957357 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.003988 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-4s45c"] Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.017436 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-4s45c"] Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.061454 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-host\") pod \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\" (UID: \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\") " Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.061500 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdq4h\" (UniqueName: \"kubernetes.io/projected/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-kube-api-access-hdq4h\") pod \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\" (UID: \"a76fb95e-5ab9-4dfb-b055-d3e0987b86a2\") " Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.061635 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-host" (OuterVolumeSpecName: "host") pod "a76fb95e-5ab9-4dfb-b055-d3e0987b86a2" (UID: "a76fb95e-5ab9-4dfb-b055-d3e0987b86a2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.062012 4935 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-host\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.068347 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-kube-api-access-hdq4h" (OuterVolumeSpecName: "kube-api-access-hdq4h") pod "a76fb95e-5ab9-4dfb-b055-d3e0987b86a2" (UID: "a76fb95e-5ab9-4dfb-b055-d3e0987b86a2"). InnerVolumeSpecName "kube-api-access-hdq4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.164330 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdq4h\" (UniqueName: \"kubernetes.io/projected/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2-kube-api-access-hdq4h\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.278011 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.366950 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-utilities\") pod \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.367152 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-catalog-content\") pod \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.367340 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n298g\" (UniqueName: \"kubernetes.io/projected/d2a84e74-15df-4e59-be99-0232bf5ffaa7-kube-api-access-n298g\") pod \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\" (UID: \"d2a84e74-15df-4e59-be99-0232bf5ffaa7\") " Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.367913 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-utilities" (OuterVolumeSpecName: "utilities") pod "d2a84e74-15df-4e59-be99-0232bf5ffaa7" (UID: "d2a84e74-15df-4e59-be99-0232bf5ffaa7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.373105 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2a84e74-15df-4e59-be99-0232bf5ffaa7-kube-api-access-n298g" (OuterVolumeSpecName: "kube-api-access-n298g") pod "d2a84e74-15df-4e59-be99-0232bf5ffaa7" (UID: "d2a84e74-15df-4e59-be99-0232bf5ffaa7"). InnerVolumeSpecName "kube-api-access-n298g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.459169 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2a84e74-15df-4e59-be99-0232bf5ffaa7" (UID: "d2a84e74-15df-4e59-be99-0232bf5ffaa7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.470331 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.470370 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n298g\" (UniqueName: \"kubernetes.io/projected/d2a84e74-15df-4e59-be99-0232bf5ffaa7-kube-api-access-n298g\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.470382 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a84e74-15df-4e59-be99-0232bf5ffaa7-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.778077 4935 generic.go:334] "Generic (PLEG): container finished" podID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerID="cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34" exitCode=0 Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.778181 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.784471 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-4s45c" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.794119 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a76fb95e-5ab9-4dfb-b055-d3e0987b86a2" path="/var/lib/kubelet/pods/a76fb95e-5ab9-4dfb-b055-d3e0987b86a2/volumes" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.795035 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv6sr" event={"ID":"d2a84e74-15df-4e59-be99-0232bf5ffaa7","Type":"ContainerDied","Data":"cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34"} Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.795069 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zv6sr" event={"ID":"d2a84e74-15df-4e59-be99-0232bf5ffaa7","Type":"ContainerDied","Data":"1c3fc9b8f8d15fd5168316ff58261072bd783c12baebcb4b6958ae273b7f7cb0"} Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.795096 4935 scope.go:117] "RemoveContainer" containerID="cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.837637 4935 scope.go:117] "RemoveContainer" containerID="6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.863224 4935 scope.go:117] "RemoveContainer" containerID="862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.904450 4935 scope.go:117] "RemoveContainer" containerID="cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34" Oct 05 10:21:40 crc kubenswrapper[4935]: E1005 10:21:40.905414 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34\": container with ID starting with cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34 not found: ID does not exist" containerID="cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.905453 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34"} err="failed to get container status \"cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34\": rpc error: code = NotFound desc = could not find container \"cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34\": container with ID starting with cf11fdb5d9a7efedbee81f7fee695940a7ffa3796917bd768b663958d9ac4c34 not found: ID does not exist" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.905478 4935 scope.go:117] "RemoveContainer" containerID="6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154" Oct 05 10:21:40 crc kubenswrapper[4935]: E1005 10:21:40.907142 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154\": container with ID starting with 6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154 not found: ID does not exist" containerID="6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.907192 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154"} err="failed to get container status \"6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154\": rpc error: code = NotFound desc = could not find container \"6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154\": container with ID starting with 6bd1a71666dd3fc6fa553c0587699f0888e5631732cd7e6ab3e3c3dde7179154 not found: ID does not exist" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.907224 4935 scope.go:117] "RemoveContainer" containerID="862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7" Oct 05 10:21:40 crc kubenswrapper[4935]: E1005 10:21:40.909363 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7\": container with ID starting with 862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7 not found: ID does not exist" containerID="862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.909437 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7"} err="failed to get container status \"862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7\": rpc error: code = NotFound desc = could not find container \"862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7\": container with ID starting with 862a7fe02bb24a295e49d50f407de425d502dde3e55507e12aaa81f4ecb53eb7 not found: ID does not exist" Oct 05 10:21:40 crc kubenswrapper[4935]: I1005 10:21:40.909482 4935 scope.go:117] "RemoveContainer" containerID="283de8d867f25e9e5ef8888a02dfeda250f160601db79293ff3ec803c47b4441" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.227426 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-n8tfd"] Oct 05 10:21:41 crc kubenswrapper[4935]: E1005 10:21:41.229776 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="extract-utilities" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.229814 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="extract-utilities" Oct 05 10:21:41 crc kubenswrapper[4935]: E1005 10:21:41.229854 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a76fb95e-5ab9-4dfb-b055-d3e0987b86a2" containerName="container-00" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.229861 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="a76fb95e-5ab9-4dfb-b055-d3e0987b86a2" containerName="container-00" Oct 05 10:21:41 crc kubenswrapper[4935]: E1005 10:21:41.229949 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerName="registry-server" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.229958 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerName="registry-server" Oct 05 10:21:41 crc kubenswrapper[4935]: E1005 10:21:41.230011 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerName="extract-utilities" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.230019 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerName="extract-utilities" Oct 05 10:21:41 crc kubenswrapper[4935]: E1005 10:21:41.230039 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="registry-server" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.230050 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="registry-server" Oct 05 10:21:41 crc kubenswrapper[4935]: E1005 10:21:41.230082 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="extract-content" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.230089 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="extract-content" Oct 05 10:21:41 crc kubenswrapper[4935]: E1005 10:21:41.230132 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerName="extract-content" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.230141 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerName="extract-content" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.230850 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="a76fb95e-5ab9-4dfb-b055-d3e0987b86a2" containerName="container-00" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.230907 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" containerName="registry-server" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.230930 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e353ad-7d3d-4500-8f35-af7eb281caab" containerName="registry-server" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.232253 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.295459 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tc7t\" (UniqueName: \"kubernetes.io/projected/2765052f-1aa0-4d14-a901-a03ed7be4c8d-kube-api-access-8tc7t\") pod \"crc-debug-n8tfd\" (UID: \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\") " pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.295658 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2765052f-1aa0-4d14-a901-a03ed7be4c8d-host\") pod \"crc-debug-n8tfd\" (UID: \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\") " pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.398235 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2765052f-1aa0-4d14-a901-a03ed7be4c8d-host\") pod \"crc-debug-n8tfd\" (UID: \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\") " pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.398349 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tc7t\" (UniqueName: \"kubernetes.io/projected/2765052f-1aa0-4d14-a901-a03ed7be4c8d-kube-api-access-8tc7t\") pod \"crc-debug-n8tfd\" (UID: \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\") " pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.398627 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2765052f-1aa0-4d14-a901-a03ed7be4c8d-host\") pod \"crc-debug-n8tfd\" (UID: \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\") " pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.420090 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tc7t\" (UniqueName: \"kubernetes.io/projected/2765052f-1aa0-4d14-a901-a03ed7be4c8d-kube-api-access-8tc7t\") pod \"crc-debug-n8tfd\" (UID: \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\") " pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.558558 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:41 crc kubenswrapper[4935]: I1005 10:21:41.801156 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" event={"ID":"2765052f-1aa0-4d14-a901-a03ed7be4c8d","Type":"ContainerStarted","Data":"5c2d669edd5bcb0d29090b332d8a2f6090c9bbc53676a21f2391066304c18535"} Oct 05 10:21:42 crc kubenswrapper[4935]: I1005 10:21:42.830569 4935 generic.go:334] "Generic (PLEG): container finished" podID="2765052f-1aa0-4d14-a901-a03ed7be4c8d" containerID="ab073a62fa6eb6a7112f3dc05222a247726e508dd89db47c4d7bd949af99495c" exitCode=0 Oct 05 10:21:42 crc kubenswrapper[4935]: I1005 10:21:42.830664 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" event={"ID":"2765052f-1aa0-4d14-a901-a03ed7be4c8d","Type":"ContainerDied","Data":"ab073a62fa6eb6a7112f3dc05222a247726e508dd89db47c4d7bd949af99495c"} Oct 05 10:21:43 crc kubenswrapper[4935]: I1005 10:21:43.941355 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.059360 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tc7t\" (UniqueName: \"kubernetes.io/projected/2765052f-1aa0-4d14-a901-a03ed7be4c8d-kube-api-access-8tc7t\") pod \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\" (UID: \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\") " Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.059485 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2765052f-1aa0-4d14-a901-a03ed7be4c8d-host\") pod \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\" (UID: \"2765052f-1aa0-4d14-a901-a03ed7be4c8d\") " Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.060614 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2765052f-1aa0-4d14-a901-a03ed7be4c8d-host" (OuterVolumeSpecName: "host") pod "2765052f-1aa0-4d14-a901-a03ed7be4c8d" (UID: "2765052f-1aa0-4d14-a901-a03ed7be4c8d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.061083 4935 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2765052f-1aa0-4d14-a901-a03ed7be4c8d-host\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.084983 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2765052f-1aa0-4d14-a901-a03ed7be4c8d-kube-api-access-8tc7t" (OuterVolumeSpecName: "kube-api-access-8tc7t") pod "2765052f-1aa0-4d14-a901-a03ed7be4c8d" (UID: "2765052f-1aa0-4d14-a901-a03ed7be4c8d"). InnerVolumeSpecName "kube-api-access-8tc7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.162864 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tc7t\" (UniqueName: \"kubernetes.io/projected/2765052f-1aa0-4d14-a901-a03ed7be4c8d-kube-api-access-8tc7t\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.886460 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" event={"ID":"2765052f-1aa0-4d14-a901-a03ed7be4c8d","Type":"ContainerDied","Data":"5c2d669edd5bcb0d29090b332d8a2f6090c9bbc53676a21f2391066304c18535"} Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.886766 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c2d669edd5bcb0d29090b332d8a2f6090c9bbc53676a21f2391066304c18535" Oct 05 10:21:44 crc kubenswrapper[4935]: I1005 10:21:44.886866 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-n8tfd" Oct 05 10:21:49 crc kubenswrapper[4935]: I1005 10:21:49.778330 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:21:49 crc kubenswrapper[4935]: E1005 10:21:49.778926 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:21:52 crc kubenswrapper[4935]: I1005 10:21:52.353009 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-n8tfd"] Oct 05 10:21:52 crc kubenswrapper[4935]: I1005 10:21:52.362188 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-n8tfd"] Oct 05 10:21:52 crc kubenswrapper[4935]: I1005 10:21:52.789238 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2765052f-1aa0-4d14-a901-a03ed7be4c8d" path="/var/lib/kubelet/pods/2765052f-1aa0-4d14-a901-a03ed7be4c8d/volumes" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.599614 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-6kwcp"] Oct 05 10:21:53 crc kubenswrapper[4935]: E1005 10:21:53.600190 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2765052f-1aa0-4d14-a901-a03ed7be4c8d" containerName="container-00" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.600205 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="2765052f-1aa0-4d14-a901-a03ed7be4c8d" containerName="container-00" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.600485 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="2765052f-1aa0-4d14-a901-a03ed7be4c8d" containerName="container-00" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.601430 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.686213 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l54r6\" (UniqueName: \"kubernetes.io/projected/b13d851c-0bf1-4048-a137-0410a72db461-kube-api-access-l54r6\") pod \"crc-debug-6kwcp\" (UID: \"b13d851c-0bf1-4048-a137-0410a72db461\") " pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.686596 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b13d851c-0bf1-4048-a137-0410a72db461-host\") pod \"crc-debug-6kwcp\" (UID: \"b13d851c-0bf1-4048-a137-0410a72db461\") " pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.788474 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l54r6\" (UniqueName: \"kubernetes.io/projected/b13d851c-0bf1-4048-a137-0410a72db461-kube-api-access-l54r6\") pod \"crc-debug-6kwcp\" (UID: \"b13d851c-0bf1-4048-a137-0410a72db461\") " pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.788621 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b13d851c-0bf1-4048-a137-0410a72db461-host\") pod \"crc-debug-6kwcp\" (UID: \"b13d851c-0bf1-4048-a137-0410a72db461\") " pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.788811 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b13d851c-0bf1-4048-a137-0410a72db461-host\") pod \"crc-debug-6kwcp\" (UID: \"b13d851c-0bf1-4048-a137-0410a72db461\") " pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.808588 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l54r6\" (UniqueName: \"kubernetes.io/projected/b13d851c-0bf1-4048-a137-0410a72db461-kube-api-access-l54r6\") pod \"crc-debug-6kwcp\" (UID: \"b13d851c-0bf1-4048-a137-0410a72db461\") " pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:53 crc kubenswrapper[4935]: I1005 10:21:53.931714 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:55 crc kubenswrapper[4935]: I1005 10:21:55.009496 4935 generic.go:334] "Generic (PLEG): container finished" podID="b13d851c-0bf1-4048-a137-0410a72db461" containerID="6ff3e9cf310ee80b7b182a504a19abd682f751d5af8381ee397fb1528cc94d2b" exitCode=0 Oct 05 10:21:55 crc kubenswrapper[4935]: I1005 10:21:55.009604 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" event={"ID":"b13d851c-0bf1-4048-a137-0410a72db461","Type":"ContainerDied","Data":"6ff3e9cf310ee80b7b182a504a19abd682f751d5af8381ee397fb1528cc94d2b"} Oct 05 10:21:55 crc kubenswrapper[4935]: I1005 10:21:55.010045 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" event={"ID":"b13d851c-0bf1-4048-a137-0410a72db461","Type":"ContainerStarted","Data":"181ac69a1f439448556ec294b9f2d6850a859eb28ffe8b1c69377932b748181e"} Oct 05 10:21:55 crc kubenswrapper[4935]: I1005 10:21:55.062546 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-6kwcp"] Oct 05 10:21:55 crc kubenswrapper[4935]: I1005 10:21:55.073625 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9m5xl/crc-debug-6kwcp"] Oct 05 10:21:56 crc kubenswrapper[4935]: I1005 10:21:56.140571 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:21:56 crc kubenswrapper[4935]: I1005 10:21:56.255318 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l54r6\" (UniqueName: \"kubernetes.io/projected/b13d851c-0bf1-4048-a137-0410a72db461-kube-api-access-l54r6\") pod \"b13d851c-0bf1-4048-a137-0410a72db461\" (UID: \"b13d851c-0bf1-4048-a137-0410a72db461\") " Oct 05 10:21:56 crc kubenswrapper[4935]: I1005 10:21:56.255505 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b13d851c-0bf1-4048-a137-0410a72db461-host\") pod \"b13d851c-0bf1-4048-a137-0410a72db461\" (UID: \"b13d851c-0bf1-4048-a137-0410a72db461\") " Oct 05 10:21:56 crc kubenswrapper[4935]: I1005 10:21:56.256118 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b13d851c-0bf1-4048-a137-0410a72db461-host" (OuterVolumeSpecName: "host") pod "b13d851c-0bf1-4048-a137-0410a72db461" (UID: "b13d851c-0bf1-4048-a137-0410a72db461"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 10:21:56 crc kubenswrapper[4935]: I1005 10:21:56.280093 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13d851c-0bf1-4048-a137-0410a72db461-kube-api-access-l54r6" (OuterVolumeSpecName: "kube-api-access-l54r6") pod "b13d851c-0bf1-4048-a137-0410a72db461" (UID: "b13d851c-0bf1-4048-a137-0410a72db461"). InnerVolumeSpecName "kube-api-access-l54r6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:21:56 crc kubenswrapper[4935]: I1005 10:21:56.358180 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l54r6\" (UniqueName: \"kubernetes.io/projected/b13d851c-0bf1-4048-a137-0410a72db461-kube-api-access-l54r6\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:56 crc kubenswrapper[4935]: I1005 10:21:56.358219 4935 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b13d851c-0bf1-4048-a137-0410a72db461-host\") on node \"crc\" DevicePath \"\"" Oct 05 10:21:56 crc kubenswrapper[4935]: I1005 10:21:56.799398 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b13d851c-0bf1-4048-a137-0410a72db461" path="/var/lib/kubelet/pods/b13d851c-0bf1-4048-a137-0410a72db461/volumes" Oct 05 10:21:57 crc kubenswrapper[4935]: I1005 10:21:57.032862 4935 scope.go:117] "RemoveContainer" containerID="6ff3e9cf310ee80b7b182a504a19abd682f751d5af8381ee397fb1528cc94d2b" Oct 05 10:21:57 crc kubenswrapper[4935]: I1005 10:21:57.032970 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/crc-debug-6kwcp" Oct 05 10:22:02 crc kubenswrapper[4935]: I1005 10:22:02.778211 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:22:02 crc kubenswrapper[4935]: E1005 10:22:02.779039 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:22:10 crc kubenswrapper[4935]: I1005 10:22:10.816451 4935 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","podd2a84e74-15df-4e59-be99-0232bf5ffaa7"] err="unable to destroy cgroup paths for cgroup [kubepods burstable podd2a84e74-15df-4e59-be99-0232bf5ffaa7] : Timed out while waiting for systemd to remove kubepods-burstable-podd2a84e74_15df_4e59_be99_0232bf5ffaa7.slice" Oct 05 10:22:10 crc kubenswrapper[4935]: E1005 10:22:10.817414 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable podd2a84e74-15df-4e59-be99-0232bf5ffaa7] : unable to destroy cgroup paths for cgroup [kubepods burstable podd2a84e74-15df-4e59-be99-0232bf5ffaa7] : Timed out while waiting for systemd to remove kubepods-burstable-podd2a84e74_15df_4e59_be99_0232bf5ffaa7.slice" pod="openshift-marketplace/redhat-operators-zv6sr" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" Oct 05 10:22:11 crc kubenswrapper[4935]: I1005 10:22:11.184988 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zv6sr" Oct 05 10:22:11 crc kubenswrapper[4935]: I1005 10:22:11.232207 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zv6sr"] Oct 05 10:22:11 crc kubenswrapper[4935]: I1005 10:22:11.241915 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zv6sr"] Oct 05 10:22:12 crc kubenswrapper[4935]: I1005 10:22:12.802080 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2a84e74-15df-4e59-be99-0232bf5ffaa7" path="/var/lib/kubelet/pods/d2a84e74-15df-4e59-be99-0232bf5ffaa7/volumes" Oct 05 10:22:14 crc kubenswrapper[4935]: I1005 10:22:14.777719 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:22:14 crc kubenswrapper[4935]: E1005 10:22:14.778396 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.239749 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v4cl4"] Oct 05 10:22:27 crc kubenswrapper[4935]: E1005 10:22:27.241063 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13d851c-0bf1-4048-a137-0410a72db461" containerName="container-00" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.241078 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13d851c-0bf1-4048-a137-0410a72db461" containerName="container-00" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.241310 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13d851c-0bf1-4048-a137-0410a72db461" containerName="container-00" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.248564 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.302935 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4cl4"] Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.368078 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8kwh\" (UniqueName: \"kubernetes.io/projected/9c51d88f-00db-488d-860f-7fed6c982492-kube-api-access-s8kwh\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.368184 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-catalog-content\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.368343 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-utilities\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.471218 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8kwh\" (UniqueName: \"kubernetes.io/projected/9c51d88f-00db-488d-860f-7fed6c982492-kube-api-access-s8kwh\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.471305 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-catalog-content\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.471483 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-utilities\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.472034 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-utilities\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.472582 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-catalog-content\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.496291 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8kwh\" (UniqueName: \"kubernetes.io/projected/9c51d88f-00db-488d-860f-7fed6c982492-kube-api-access-s8kwh\") pod \"redhat-marketplace-v4cl4\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:27 crc kubenswrapper[4935]: I1005 10:22:27.607085 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:28 crc kubenswrapper[4935]: I1005 10:22:28.139755 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4cl4"] Oct 05 10:22:28 crc kubenswrapper[4935]: I1005 10:22:28.393745 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4cl4" event={"ID":"9c51d88f-00db-488d-860f-7fed6c982492","Type":"ContainerStarted","Data":"1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8"} Oct 05 10:22:28 crc kubenswrapper[4935]: I1005 10:22:28.394109 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4cl4" event={"ID":"9c51d88f-00db-488d-860f-7fed6c982492","Type":"ContainerStarted","Data":"f5c402d25231edc972371690989ae84899cf3c5d8cbd393c86a47ac803ecba9c"} Oct 05 10:22:28 crc kubenswrapper[4935]: I1005 10:22:28.778062 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:22:28 crc kubenswrapper[4935]: E1005 10:22:28.778339 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:22:29 crc kubenswrapper[4935]: I1005 10:22:29.414631 4935 generic.go:334] "Generic (PLEG): container finished" podID="9c51d88f-00db-488d-860f-7fed6c982492" containerID="1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8" exitCode=0 Oct 05 10:22:29 crc kubenswrapper[4935]: I1005 10:22:29.415106 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4cl4" event={"ID":"9c51d88f-00db-488d-860f-7fed6c982492","Type":"ContainerDied","Data":"1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8"} Oct 05 10:22:29 crc kubenswrapper[4935]: I1005 10:22:29.415629 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4cl4" event={"ID":"9c51d88f-00db-488d-860f-7fed6c982492","Type":"ContainerStarted","Data":"a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707"} Oct 05 10:22:30 crc kubenswrapper[4935]: I1005 10:22:30.430317 4935 generic.go:334] "Generic (PLEG): container finished" podID="9c51d88f-00db-488d-860f-7fed6c982492" containerID="a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707" exitCode=0 Oct 05 10:22:30 crc kubenswrapper[4935]: I1005 10:22:30.430455 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4cl4" event={"ID":"9c51d88f-00db-488d-860f-7fed6c982492","Type":"ContainerDied","Data":"a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707"} Oct 05 10:22:31 crc kubenswrapper[4935]: I1005 10:22:31.450164 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4cl4" event={"ID":"9c51d88f-00db-488d-860f-7fed6c982492","Type":"ContainerStarted","Data":"5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5"} Oct 05 10:22:31 crc kubenswrapper[4935]: I1005 10:22:31.498485 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v4cl4" podStartSLOduration=2.090230113 podStartE2EDuration="4.49845448s" podCreationTimestamp="2025-10-05 10:22:27 +0000 UTC" firstStartedPulling="2025-10-05 10:22:28.39705457 +0000 UTC m=+12582.279681020" lastFinishedPulling="2025-10-05 10:22:30.805278917 +0000 UTC m=+12584.687905387" observedRunningTime="2025-10-05 10:22:31.4852274 +0000 UTC m=+12585.367853890" watchObservedRunningTime="2025-10-05 10:22:31.49845448 +0000 UTC m=+12585.381080940" Oct 05 10:22:37 crc kubenswrapper[4935]: I1005 10:22:37.609216 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:37 crc kubenswrapper[4935]: I1005 10:22:37.610255 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:37 crc kubenswrapper[4935]: I1005 10:22:37.703259 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:38 crc kubenswrapper[4935]: I1005 10:22:38.625952 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:38 crc kubenswrapper[4935]: I1005 10:22:38.682018 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4cl4"] Oct 05 10:22:40 crc kubenswrapper[4935]: I1005 10:22:40.575325 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v4cl4" podUID="9c51d88f-00db-488d-860f-7fed6c982492" containerName="registry-server" containerID="cri-o://5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5" gracePeriod=2 Oct 05 10:22:40 crc kubenswrapper[4935]: I1005 10:22:40.976325 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_f5d04268-2dab-4d23-b434-b91cb01108f9/init-config-reloader/0.log" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.041315 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.098160 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8kwh\" (UniqueName: \"kubernetes.io/projected/9c51d88f-00db-488d-860f-7fed6c982492-kube-api-access-s8kwh\") pod \"9c51d88f-00db-488d-860f-7fed6c982492\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.098233 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-catalog-content\") pod \"9c51d88f-00db-488d-860f-7fed6c982492\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.098636 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-utilities\") pod \"9c51d88f-00db-488d-860f-7fed6c982492\" (UID: \"9c51d88f-00db-488d-860f-7fed6c982492\") " Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.100485 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-utilities" (OuterVolumeSpecName: "utilities") pod "9c51d88f-00db-488d-860f-7fed6c982492" (UID: "9c51d88f-00db-488d-860f-7fed6c982492"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.106180 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c51d88f-00db-488d-860f-7fed6c982492-kube-api-access-s8kwh" (OuterVolumeSpecName: "kube-api-access-s8kwh") pod "9c51d88f-00db-488d-860f-7fed6c982492" (UID: "9c51d88f-00db-488d-860f-7fed6c982492"). InnerVolumeSpecName "kube-api-access-s8kwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.118093 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c51d88f-00db-488d-860f-7fed6c982492" (UID: "9c51d88f-00db-488d-860f-7fed6c982492"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.143446 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_f5d04268-2dab-4d23-b434-b91cb01108f9/init-config-reloader/0.log" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.153673 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_f5d04268-2dab-4d23-b434-b91cb01108f9/alertmanager/0.log" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.200778 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.200823 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8kwh\" (UniqueName: \"kubernetes.io/projected/9c51d88f-00db-488d-860f-7fed6c982492-kube-api-access-s8kwh\") on node \"crc\" DevicePath \"\"" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.200832 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c51d88f-00db-488d-860f-7fed6c982492-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.287325 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_f5d04268-2dab-4d23-b434-b91cb01108f9/config-reloader/0.log" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.449933 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_b6e988f2-3f23-4152-9830-5e03d950de38/aodh-api/0.log" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.550878 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_b6e988f2-3f23-4152-9830-5e03d950de38/aodh-evaluator/0.log" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.585530 4935 generic.go:334] "Generic (PLEG): container finished" podID="9c51d88f-00db-488d-860f-7fed6c982492" containerID="5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5" exitCode=0 Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.585582 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4cl4" event={"ID":"9c51d88f-00db-488d-860f-7fed6c982492","Type":"ContainerDied","Data":"5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5"} Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.585615 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v4cl4" event={"ID":"9c51d88f-00db-488d-860f-7fed6c982492","Type":"ContainerDied","Data":"f5c402d25231edc972371690989ae84899cf3c5d8cbd393c86a47ac803ecba9c"} Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.585635 4935 scope.go:117] "RemoveContainer" containerID="5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.585802 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v4cl4" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.612940 4935 scope.go:117] "RemoveContainer" containerID="a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.629383 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4cl4"] Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.634470 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v4cl4"] Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.638840 4935 scope.go:117] "RemoveContainer" containerID="1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.680325 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_b6e988f2-3f23-4152-9830-5e03d950de38/aodh-listener/0.log" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.685279 4935 scope.go:117] "RemoveContainer" containerID="5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5" Oct 05 10:22:41 crc kubenswrapper[4935]: E1005 10:22:41.685941 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5\": container with ID starting with 5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5 not found: ID does not exist" containerID="5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.685975 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5"} err="failed to get container status \"5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5\": rpc error: code = NotFound desc = could not find container \"5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5\": container with ID starting with 5353bf921c262ef95415e1ed94c2065f2881c286d81ce2879ffab4d64c5d9cd5 not found: ID does not exist" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.685994 4935 scope.go:117] "RemoveContainer" containerID="a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707" Oct 05 10:22:41 crc kubenswrapper[4935]: E1005 10:22:41.691034 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707\": container with ID starting with a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707 not found: ID does not exist" containerID="a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.691061 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707"} err="failed to get container status \"a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707\": rpc error: code = NotFound desc = could not find container \"a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707\": container with ID starting with a4a9545bb704059842047796deb22fd3d336d9db80574869679dfceb9f6a5707 not found: ID does not exist" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.691074 4935 scope.go:117] "RemoveContainer" containerID="1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8" Oct 05 10:22:41 crc kubenswrapper[4935]: E1005 10:22:41.694017 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8\": container with ID starting with 1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8 not found: ID does not exist" containerID="1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.694054 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8"} err="failed to get container status \"1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8\": rpc error: code = NotFound desc = could not find container \"1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8\": container with ID starting with 1560e42af88298c7bdbae99681fef41dca84ae2b7e2100105e3ad01803c5e1d8 not found: ID does not exist" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.761460 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_b6e988f2-3f23-4152-9830-5e03d950de38/aodh-notifier/0.log" Oct 05 10:22:41 crc kubenswrapper[4935]: I1005 10:22:41.920186 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-85d576456b-gxmw2_dd1c0eed-cccf-453d-8e01-59ce9890f7eb/barbican-api/0.log" Oct 05 10:22:42 crc kubenswrapper[4935]: I1005 10:22:42.084595 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-85d576456b-gxmw2_dd1c0eed-cccf-453d-8e01-59ce9890f7eb/barbican-api-log/0.log" Oct 05 10:22:42 crc kubenswrapper[4935]: I1005 10:22:42.308989 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-57cf597c6d-gng7h_e4c4f39e-ff2a-48a2-acb5-df5542a6e218/barbican-keystone-listener/0.log" Oct 05 10:22:42 crc kubenswrapper[4935]: I1005 10:22:42.789256 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c51d88f-00db-488d-860f-7fed6c982492" path="/var/lib/kubelet/pods/9c51d88f-00db-488d-860f-7fed6c982492/volumes" Oct 05 10:22:42 crc kubenswrapper[4935]: I1005 10:22:42.813936 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c558b76b7-7vtbc_dbb4faf3-8787-4ca3-a956-6794a9ad86ea/barbican-worker/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.016317 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c558b76b7-7vtbc_dbb4faf3-8787-4ca3-a956-6794a9ad86ea/barbican-worker-log/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.183171 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-57cf597c6d-gng7h_e4c4f39e-ff2a-48a2-acb5-df5542a6e218/barbican-keystone-listener-log/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.238089 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-rtkmw_3da548b8-5970-4e0c-9b11-261950cb6e8d/bootstrap-openstack-openstack-cell1/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.434677 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-networker-bk27c_6870738c-a0fd-4448-b129-be1246e47542/bootstrap-openstack-openstack-networker/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.519571 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_be33a62f-6a3c-4893-8a04-c67b5c08159f/ceilometer-central-agent/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.612410 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_be33a62f-6a3c-4893-8a04-c67b5c08159f/ceilometer-notification-agent/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.654887 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_be33a62f-6a3c-4893-8a04-c67b5c08159f/proxy-httpd/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.706532 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_be33a62f-6a3c-4893-8a04-c67b5c08159f/sg-core/0.log" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.777948 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:22:43 crc kubenswrapper[4935]: E1005 10:22:43.778328 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:22:43 crc kubenswrapper[4935]: I1005 10:22:43.813111 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-gt9nr_ecda3b70-058d-490d-a537-7511163dad56/ceph-client-openstack-openstack-cell1/0.log" Oct 05 10:22:44 crc kubenswrapper[4935]: I1005 10:22:44.399424 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9cec9943-9a6f-4752-b334-f49d622c1929/cinder-api-log/0.log" Oct 05 10:22:44 crc kubenswrapper[4935]: I1005 10:22:44.473620 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9cec9943-9a6f-4752-b334-f49d622c1929/cinder-api/0.log" Oct 05 10:22:44 crc kubenswrapper[4935]: I1005 10:22:44.646161 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_7cb2a652-0e57-4699-85d5-69e6966e8948/probe/0.log" Oct 05 10:22:44 crc kubenswrapper[4935]: I1005 10:22:44.865620 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2ae0b7bf-20f9-4942-898f-d44b8c061d30/cinder-scheduler/0.log" Oct 05 10:22:44 crc kubenswrapper[4935]: I1005 10:22:44.959532 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2ae0b7bf-20f9-4942-898f-d44b8c061d30/probe/0.log" Oct 05 10:22:45 crc kubenswrapper[4935]: I1005 10:22:45.410768 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_490ba015-eba1-4f8e-a935-f8d27de1dc8d/probe/0.log" Oct 05 10:22:45 crc kubenswrapper[4935]: I1005 10:22:45.832249 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-6j8z7_a90bd676-a895-44f5-b3b5-c23a9178c9e0/configure-network-openstack-openstack-cell1/0.log" Oct 05 10:22:46 crc kubenswrapper[4935]: I1005 10:22:46.253414 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-networker-wq6jl_2bec0abb-c7e2-4a0b-8075-83158bcd560f/configure-network-openstack-openstack-networker/0.log" Oct 05 10:22:46 crc kubenswrapper[4935]: I1005 10:22:46.722425 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-4cfll_697225c9-e0f0-46b7-9fc1-6456384033bb/configure-os-openstack-openstack-cell1/0.log" Oct 05 10:22:47 crc kubenswrapper[4935]: I1005 10:22:47.170903 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-cn2bk_141a4fc2-2a6a-4464-bd8a-9ad826abba58/configure-os-openstack-openstack-cell1/0.log" Oct 05 10:22:47 crc kubenswrapper[4935]: I1005 10:22:47.624935 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-networker-4cgvb_abe126de-f191-4efd-b1cc-dfaa5937bc77/configure-os-openstack-openstack-networker/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.043645 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_490ba015-eba1-4f8e-a935-f8d27de1dc8d/cinder-volume/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.073766 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-networker-7nmjq_f8b0de95-9148-4f8d-aab4-997b20290690/configure-os-openstack-openstack-networker/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.170409 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_7cb2a652-0e57-4699-85d5-69e6966e8948/cinder-backup/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.272164 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7dbb57f589-r2bjx_37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e/init/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.445695 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7dbb57f589-r2bjx_37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e/init/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.527790 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7dbb57f589-r2bjx_37ba0167-1ffc-4a1f-8b11-3dc0d0f15c8e/dnsmasq-dns/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.629137 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-ssp86_e39fcf7d-8ea6-4a0a-87ed-a6000793d030/download-cache-openstack-openstack-cell1/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.711794 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-networker-blbmn_6d1fcc45-8bfe-4406-ab6d-fa30a85985cc/download-cache-openstack-openstack-networker/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.909667 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_59119f30-1499-4892-a9f8-997e38750ded/glance-log/0.log" Oct 05 10:22:48 crc kubenswrapper[4935]: I1005 10:22:48.930559 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_59119f30-1499-4892-a9f8-997e38750ded/glance-httpd/0.log" Oct 05 10:22:49 crc kubenswrapper[4935]: I1005 10:22:49.095507 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ef490248-2d23-4834-8a80-334916da488f/glance-log/0.log" Oct 05 10:22:49 crc kubenswrapper[4935]: I1005 10:22:49.135240 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ef490248-2d23-4834-8a80-334916da488f/glance-httpd/0.log" Oct 05 10:22:49 crc kubenswrapper[4935]: I1005 10:22:49.436386 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-54d757dd4f-264jb_5b2d6e9e-ae50-4ee7-adc7-2ce597e40d17/heat-api/0.log" Oct 05 10:22:49 crc kubenswrapper[4935]: I1005 10:22:49.590522 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-59984465f-5qgvh_dfeb90b7-9abc-4f02-8a73-2c6df751c902/heat-cfnapi/0.log" Oct 05 10:22:49 crc kubenswrapper[4935]: I1005 10:22:49.661635 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-59d5d97855-8b85w_f3cb9fa5-3711-4f0a-830b-e2d210caada6/heat-engine/0.log" Oct 05 10:22:49 crc kubenswrapper[4935]: I1005 10:22:49.951962 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-67b8b7d97c-2hshb_c15807dd-5d29-4c60-a99e-c80dd38c0b27/horizon/0.log" Oct 05 10:22:49 crc kubenswrapper[4935]: I1005 10:22:49.985827 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-67b8b7d97c-2hshb_c15807dd-5d29-4c60-a99e-c80dd38c0b27/horizon-log/0.log" Oct 05 10:22:50 crc kubenswrapper[4935]: I1005 10:22:50.158841 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-vpc8x_f9f3edd1-359c-4a45-bcdb-026066428d7f/install-certs-openstack-openstack-cell1/0.log" Oct 05 10:22:50 crc kubenswrapper[4935]: I1005 10:22:50.210291 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-networker-5rb6l_6efd4306-da0b-4596-a2b8-20614f208aa7/install-certs-openstack-openstack-networker/0.log" Oct 05 10:22:50 crc kubenswrapper[4935]: I1005 10:22:50.397594 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-bqz75_9113dc85-b57c-447f-ab4d-1294dcd0f61f/install-os-openstack-openstack-cell1/0.log" Oct 05 10:22:50 crc kubenswrapper[4935]: I1005 10:22:50.569394 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-networker-xkr98_cdbef832-881e-4a2a-8735-59f801251dd5/install-os-openstack-openstack-networker/0.log" Oct 05 10:22:50 crc kubenswrapper[4935]: I1005 10:22:50.665406 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29327581-lgtqk_57010d7c-9499-4337-b1de-29766c39327a/keystone-cron/0.log" Oct 05 10:22:50 crc kubenswrapper[4935]: I1005 10:22:50.859539 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29327641-kw7hz_f004cad9-5506-4c21-8fbc-9bc72bb3470a/keystone-cron/0.log" Oct 05 10:22:51 crc kubenswrapper[4935]: I1005 10:22:51.131216 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_71af6336-7272-441b-ba9e-a14b824af3c4/kube-state-metrics/0.log" Oct 05 10:22:51 crc kubenswrapper[4935]: I1005 10:22:51.416212 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-fslpj_4eec0ed5-4af6-45bd-9196-e3ebd559575b/libvirt-openstack-openstack-cell1/0.log" Oct 05 10:22:52 crc kubenswrapper[4935]: I1005 10:22:52.001162 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_60f96a82-8665-49b1-9f4e-8a3923e6d996/manila-api/0.log" Oct 05 10:22:52 crc kubenswrapper[4935]: I1005 10:22:52.301171 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_60f96a82-8665-49b1-9f4e-8a3923e6d996/manila-api-log/0.log" Oct 05 10:22:52 crc kubenswrapper[4935]: I1005 10:22:52.368615 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-f8f877d-mhddp_3c556c2f-6744-4691-b651-e5359f1c8d78/keystone-api/0.log" Oct 05 10:22:52 crc kubenswrapper[4935]: I1005 10:22:52.485822 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_4340e44d-1160-461e-859a-efebef2f545b/manila-scheduler/0.log" Oct 05 10:22:52 crc kubenswrapper[4935]: I1005 10:22:52.488089 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_4340e44d-1160-461e-859a-efebef2f545b/probe/0.log" Oct 05 10:22:52 crc kubenswrapper[4935]: I1005 10:22:52.686860 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a3869bdd-ef72-488e-8691-d4c6bc8d8118/manila-share/0.log" Oct 05 10:22:52 crc kubenswrapper[4935]: I1005 10:22:52.696221 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_a3869bdd-ef72-488e-8691-d4c6bc8d8118/probe/0.log" Oct 05 10:22:53 crc kubenswrapper[4935]: I1005 10:22:53.452717 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79b64455c-9bh2f_6c373000-c6d0-4a7d-9623-2804446b1507/neutron-httpd/0.log" Oct 05 10:22:53 crc kubenswrapper[4935]: I1005 10:22:53.787136 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79b64455c-9bh2f_6c373000-c6d0-4a7d-9623-2804446b1507/neutron-api/0.log" Oct 05 10:22:53 crc kubenswrapper[4935]: I1005 10:22:53.953617 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-7qr5w_5751e378-39fa-4bf1-98b1-b1242d642c2c/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 05 10:22:54 crc kubenswrapper[4935]: I1005 10:22:54.207311 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-65rhl_c01b53f0-a632-4a5e-bd1c-cae6de7729f3/neutron-metadata-openstack-openstack-cell1/0.log" Oct 05 10:22:54 crc kubenswrapper[4935]: I1005 10:22:54.657697 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-networker-kr2ff_e903bbab-dcb0-4fc8-9265-195010d8f46e/neutron-metadata-openstack-openstack-networker/0.log" Oct 05 10:22:54 crc kubenswrapper[4935]: I1005 10:22:54.821638 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-qqmlf_fcf454be-3f70-46c3-904b-dd763a1b35a7/neutron-sriov-openstack-openstack-cell1/0.log" Oct 05 10:22:55 crc kubenswrapper[4935]: I1005 10:22:55.218277 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c69aa404-fec4-4ba7-bafe-30c254ac2d38/nova-api-api/0.log" Oct 05 10:22:55 crc kubenswrapper[4935]: I1005 10:22:55.426293 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c69aa404-fec4-4ba7-bafe-30c254ac2d38/nova-api-log/0.log" Oct 05 10:22:55 crc kubenswrapper[4935]: I1005 10:22:55.698320 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e91bdcc1-916c-432a-8df0-ee70b173a847/nova-cell0-conductor-conductor/0.log" Oct 05 10:22:56 crc kubenswrapper[4935]: I1005 10:22:56.006108 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7d7334cb-9a45-4bfa-8ec5-0d68b5b612ee/nova-cell1-conductor-conductor/0.log" Oct 05 10:22:56 crc kubenswrapper[4935]: I1005 10:22:56.223496 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_7b31e6d4-4a6a-4d28-b44f-26b72cdfd3f1/nova-cell1-novncproxy-novncproxy/0.log" Oct 05 10:22:56 crc kubenswrapper[4935]: I1005 10:22:56.568271 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celld2mv4_8a8cc4c7-3c2c-4f19-9224-8d06432ca678/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 05 10:22:56 crc kubenswrapper[4935]: I1005 10:22:56.804515 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-mrgrf_2bad5788-676a-436c-bfad-20d810bbd73e/nova-cell1-openstack-openstack-cell1/0.log" Oct 05 10:22:57 crc kubenswrapper[4935]: I1005 10:22:57.111568 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e6cc20f0-262c-4d74-b562-72e20576637f/nova-metadata-log/0.log" Oct 05 10:22:57 crc kubenswrapper[4935]: I1005 10:22:57.152072 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e6cc20f0-262c-4d74-b562-72e20576637f/nova-metadata-metadata/0.log" Oct 05 10:22:57 crc kubenswrapper[4935]: I1005 10:22:57.523690 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_afdb2053-f3c5-4665-a7c7-80f708039ae1/nova-scheduler-scheduler/0.log" Oct 05 10:22:57 crc kubenswrapper[4935]: I1005 10:22:57.739862 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d558bd5-83a1-4c77-b06e-9c499e2e49b6/mysql-bootstrap/0.log" Oct 05 10:22:57 crc kubenswrapper[4935]: I1005 10:22:57.776803 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:22:57 crc kubenswrapper[4935]: E1005 10:22:57.777388 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:22:57 crc kubenswrapper[4935]: I1005 10:22:57.892657 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d558bd5-83a1-4c77-b06e-9c499e2e49b6/mysql-bootstrap/0.log" Oct 05 10:22:57 crc kubenswrapper[4935]: I1005 10:22:57.949223 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d558bd5-83a1-4c77-b06e-9c499e2e49b6/galera/0.log" Oct 05 10:22:58 crc kubenswrapper[4935]: I1005 10:22:58.165204 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae622d60-4705-4af0-a65b-4c6654f825c7/mysql-bootstrap/0.log" Oct 05 10:22:58 crc kubenswrapper[4935]: I1005 10:22:58.347159 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae622d60-4705-4af0-a65b-4c6654f825c7/mysql-bootstrap/0.log" Oct 05 10:22:58 crc kubenswrapper[4935]: I1005 10:22:58.402564 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae622d60-4705-4af0-a65b-4c6654f825c7/galera/0.log" Oct 05 10:22:58 crc kubenswrapper[4935]: I1005 10:22:58.577922 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_e58adf26-aff2-4a22-a7bb-47fb6664d949/openstackclient/0.log" Oct 05 10:22:58 crc kubenswrapper[4935]: I1005 10:22:58.772475 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5/openstack-network-exporter/0.log" Oct 05 10:22:58 crc kubenswrapper[4935]: I1005 10:22:58.956302 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ef6b6b47-12d1-4e7a-9f0c-ab654c363fd5/ovn-northd/0.log" Oct 05 10:22:59 crc kubenswrapper[4935]: I1005 10:22:59.200411 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-lfkb6_48e06b46-b928-49eb-a8b8-5b4b5d24925b/ovn-openstack-openstack-cell1/0.log" Oct 05 10:22:59 crc kubenswrapper[4935]: I1005 10:22:59.632821 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-networker-zgwf7_7b5b8e58-6662-4dac-8700-f324b232653b/ovn-openstack-openstack-networker/0.log" Oct 05 10:22:59 crc kubenswrapper[4935]: I1005 10:22:59.686436 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_63ebf5e7-d07d-40af-890c-ecb794b3749c/openstack-network-exporter/0.log" Oct 05 10:22:59 crc kubenswrapper[4935]: I1005 10:22:59.850639 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_63ebf5e7-d07d-40af-890c-ecb794b3749c/ovsdbserver-nb/0.log" Oct 05 10:23:00 crc kubenswrapper[4935]: I1005 10:23:00.016804 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_3293245e-b8f2-4081-a526-227a9918dbe0/openstack-network-exporter/0.log" Oct 05 10:23:00 crc kubenswrapper[4935]: I1005 10:23:00.062598 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_3293245e-b8f2-4081-a526-227a9918dbe0/ovsdbserver-nb/0.log" Oct 05 10:23:00 crc kubenswrapper[4935]: I1005 10:23:00.222716 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_a258f8aa-2ec9-4098-b98f-23051ca75c08/openstack-network-exporter/0.log" Oct 05 10:23:00 crc kubenswrapper[4935]: I1005 10:23:00.390402 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_a258f8aa-2ec9-4098-b98f-23051ca75c08/ovsdbserver-nb/0.log" Oct 05 10:23:00 crc kubenswrapper[4935]: I1005 10:23:00.586646 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_419c6823-dbe7-4671-a7ea-ea7c91dff77d/openstack-network-exporter/0.log" Oct 05 10:23:00 crc kubenswrapper[4935]: I1005 10:23:00.675957 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_419c6823-dbe7-4671-a7ea-ea7c91dff77d/ovsdbserver-sb/0.log" Oct 05 10:23:00 crc kubenswrapper[4935]: I1005 10:23:00.894495 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_4315a8d6-2d0a-4576-be77-30e0b2096f0f/openstack-network-exporter/0.log" Oct 05 10:23:00 crc kubenswrapper[4935]: I1005 10:23:00.983845 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_4315a8d6-2d0a-4576-be77-30e0b2096f0f/ovsdbserver-sb/0.log" Oct 05 10:23:01 crc kubenswrapper[4935]: I1005 10:23:01.179459 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_94cd3f0f-4346-424d-9ea9-4d8153d7d418/openstack-network-exporter/0.log" Oct 05 10:23:01 crc kubenswrapper[4935]: I1005 10:23:01.301744 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_94cd3f0f-4346-424d-9ea9-4d8153d7d418/ovsdbserver-sb/0.log" Oct 05 10:23:01 crc kubenswrapper[4935]: I1005 10:23:01.760636 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7bd4bdf8fb-4xbs4_69c4aba6-a024-4c0a-b8de-213c6119bb73/placement-api/0.log" Oct 05 10:23:01 crc kubenswrapper[4935]: I1005 10:23:01.992646 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7bd4bdf8fb-4xbs4_69c4aba6-a024-4c0a-b8de-213c6119bb73/placement-log/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.145115 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_aa569082-01a4-4f8c-97a1-665f630a6e53/memcached/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.207429 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-c7hrv4_7e2ad6c2-e6d7-4592-ab04-0b13b92be5da/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.348645 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-nrsbcb_a0dab6e1-6403-4221-820a-eb5397c8f884/pre-adoption-validation-openstack-pre-adoption-openstack-networ/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.443996 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e18b6457-4bba-40d7-abad-577c4d251f45/init-config-reloader/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.568051 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e18b6457-4bba-40d7-abad-577c4d251f45/init-config-reloader/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.598915 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e18b6457-4bba-40d7-abad-577c4d251f45/config-reloader/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.640602 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e18b6457-4bba-40d7-abad-577c4d251f45/prometheus/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.742452 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e18b6457-4bba-40d7-abad-577c4d251f45/thanos-sidecar/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.821657 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0a21d4cb-88f1-4bec-ad29-97e503d2dad5/setup-container/0.log" Oct 05 10:23:02 crc kubenswrapper[4935]: I1005 10:23:02.992024 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0a21d4cb-88f1-4bec-ad29-97e503d2dad5/setup-container/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.019940 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0a21d4cb-88f1-4bec-ad29-97e503d2dad5/rabbitmq/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.147448 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377/setup-container/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.293360 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377/setup-container/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.341881 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f7d76a16-e4d9-4e84-81a8-fc4bfdfc0377/rabbitmq/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.461654 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-dd458_db79fc8f-401b-426f-9f03-aada57d9d905/reboot-os-openstack-openstack-cell1/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.506473 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-networker-dt6wh_8871439a-8466-4285-9872-747cfb0eb998/reboot-os-openstack-openstack-networker/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.677093 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-tkmbh_0982ce01-4f77-4cf8-a024-bdb05e7016f7/run-os-openstack-openstack-cell1/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.829135 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-networker-f2sdw_43c59bed-43ce-4cef-9d35-45e7496c92f3/run-os-openstack-openstack-networker/0.log" Oct 05 10:23:03 crc kubenswrapper[4935]: I1005 10:23:03.915532 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-jn258_06081368-97c9-41f4-9fed-75c593994b11/ssh-known-hosts-openstack/0.log" Oct 05 10:23:04 crc kubenswrapper[4935]: I1005 10:23:04.110298 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-c2gcm_5603bea1-f3c6-42b1-9f3c-5f4f10b4324a/telemetry-openstack-openstack-cell1/0.log" Oct 05 10:23:04 crc kubenswrapper[4935]: I1005 10:23:04.342130 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_0146f93c-5183-4802-aba2-0780ca2411c2/test-operator-logs-container/0.log" Oct 05 10:23:04 crc kubenswrapper[4935]: I1005 10:23:04.347646 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c54822d6-7348-4d81-8681-f338af0badc1/tempest-tests-tempest-tests-runner/0.log" Oct 05 10:23:04 crc kubenswrapper[4935]: I1005 10:23:04.559198 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-slgwx_097a7b90-7e36-4aa4-b9df-e17362922f55/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 05 10:23:04 crc kubenswrapper[4935]: I1005 10:23:04.858144 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-6nl7k_ed429030-8625-49b1-8e3f-106b40ae136a/validate-network-openstack-openstack-cell1/0.log" Oct 05 10:23:04 crc kubenswrapper[4935]: I1005 10:23:04.866380 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-networker-h5bcp_aad90193-5bba-4acd-ac25-d67d21a165d3/tripleo-cleanup-tripleo-cleanup-openstack-networker/0.log" Oct 05 10:23:05 crc kubenswrapper[4935]: I1005 10:23:05.014284 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-networker-cchlw_9ea0d044-3609-43f6-af64-15f5901287ad/validate-network-openstack-openstack-networker/0.log" Oct 05 10:23:11 crc kubenswrapper[4935]: I1005 10:23:11.778291 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:23:11 crc kubenswrapper[4935]: E1005 10:23:11.779054 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:23:26 crc kubenswrapper[4935]: I1005 10:23:26.785613 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:23:26 crc kubenswrapper[4935]: E1005 10:23:26.786525 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:23:28 crc kubenswrapper[4935]: I1005 10:23:28.476245 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm_fa8536f3-2840-4264-9e20-df1499d52611/util/0.log" Oct 05 10:23:28 crc kubenswrapper[4935]: I1005 10:23:28.739831 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm_fa8536f3-2840-4264-9e20-df1499d52611/pull/0.log" Oct 05 10:23:28 crc kubenswrapper[4935]: I1005 10:23:28.752918 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm_fa8536f3-2840-4264-9e20-df1499d52611/pull/0.log" Oct 05 10:23:28 crc kubenswrapper[4935]: I1005 10:23:28.774829 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm_fa8536f3-2840-4264-9e20-df1499d52611/util/0.log" Oct 05 10:23:28 crc kubenswrapper[4935]: I1005 10:23:28.989498 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm_fa8536f3-2840-4264-9e20-df1499d52611/extract/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.025744 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm_fa8536f3-2840-4264-9e20-df1499d52611/util/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.026239 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8wmkjm_fa8536f3-2840-4264-9e20-df1499d52611/pull/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.296656 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5b974f6766-6nxjj_451246ff-f04c-4b5b-a956-41592f58ddb9/manager/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.310736 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5b974f6766-6nxjj_451246ff-f04c-4b5b-a956-41592f58ddb9/kube-rbac-proxy/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.325646 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-nrs2t_1ef01eb4-46b6-4859-a53f-1b9b58f9f408/kube-rbac-proxy/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.553712 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-ntjg7_96b524ed-4a5d-4c95-b076-c7f68fab13a2/kube-rbac-proxy/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.556287 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-ntjg7_96b524ed-4a5d-4c95-b076-c7f68fab13a2/manager/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.604657 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-nrs2t_1ef01eb4-46b6-4859-a53f-1b9b58f9f408/manager/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.761305 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-698456cdc6-55r29_1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e/kube-rbac-proxy/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.893234 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-698456cdc6-55r29_1a1e7cb2-e58a-44ca-a1dc-9381d0f77c6e/manager/0.log" Oct 05 10:23:29 crc kubenswrapper[4935]: I1005 10:23:29.963219 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5c497dbdb-s8gzx_c7430e4b-67b4-45b5-8444-1ee6cf758c5b/kube-rbac-proxy/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.022107 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5c497dbdb-s8gzx_c7430e4b-67b4-45b5-8444-1ee6cf758c5b/manager/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.157515 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6675647785-zk5fs_317d3174-23c3-4af7-b289-26844e8d3f12/kube-rbac-proxy/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.190683 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6675647785-zk5fs_317d3174-23c3-4af7-b289-26844e8d3f12/manager/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.284326 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-b285k_6c0af6f9-3f50-4620-94a9-9c9262d73fd9/kube-rbac-proxy/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.549288 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f5894c49f-g84c7_eb1aa07b-d43b-4ad3-803c-042de10feab8/kube-rbac-proxy/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.550699 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f5894c49f-g84c7_eb1aa07b-d43b-4ad3-803c-042de10feab8/manager/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.563682 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-b285k_6c0af6f9-3f50-4620-94a9-9c9262d73fd9/manager/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.759248 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-57c9cdcf57-22j5g_dd907a5c-8a2c-459a-b4e0-2f1341fe956b/kube-rbac-proxy/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.860791 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-57c9cdcf57-22j5g_dd907a5c-8a2c-459a-b4e0-2f1341fe956b/manager/0.log" Oct 05 10:23:30 crc kubenswrapper[4935]: I1005 10:23:30.871550 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-lhzb8_3078f420-c252-4650-ab0a-17c92ab8d1c6/kube-rbac-proxy/0.log" Oct 05 10:23:31 crc kubenswrapper[4935]: I1005 10:23:31.026759 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-lhzb8_3078f420-c252-4650-ab0a-17c92ab8d1c6/manager/0.log" Oct 05 10:23:31 crc kubenswrapper[4935]: I1005 10:23:31.049952 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-llst6_1fb98234-7f3b-491f-871a-106ce0f2af2a/kube-rbac-proxy/0.log" Oct 05 10:23:31 crc kubenswrapper[4935]: I1005 10:23:31.163908 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-llst6_1fb98234-7f3b-491f-871a-106ce0f2af2a/manager/0.log" Oct 05 10:23:31 crc kubenswrapper[4935]: I1005 10:23:31.316881 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-55k65_5b24bc52-99be-499c-90ac-91114405b8f6/kube-rbac-proxy/0.log" Oct 05 10:23:31 crc kubenswrapper[4935]: I1005 10:23:31.355184 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-55k65_5b24bc52-99be-499c-90ac-91114405b8f6/manager/0.log" Oct 05 10:23:31 crc kubenswrapper[4935]: I1005 10:23:31.725222 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-l9bf8_1201b72a-0267-4e34-bdf5-6ddb0f9299dd/kube-rbac-proxy/0.log" Oct 05 10:23:31 crc kubenswrapper[4935]: I1005 10:23:31.789853 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-d2l5q_39f750d8-d3c5-4164-ab98-12c031e752ab/kube-rbac-proxy/0.log" Oct 05 10:23:31 crc kubenswrapper[4935]: I1005 10:23:31.948944 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-d2l5q_39f750d8-d3c5-4164-ab98-12c031e752ab/manager/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.009439 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-l9bf8_1201b72a-0267-4e34-bdf5-6ddb0f9299dd/manager/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.070409 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz_1a18b889-577c-4bb0-9d34-b2e560318144/kube-rbac-proxy/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.170454 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-66cc85b5d5tlwzz_1a18b889-577c-4bb0-9d34-b2e560318144/manager/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.347019 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7cfc658b9-vckkj_3ca2bfab-d801-4cbf-bb97-c0eff17db807/kube-rbac-proxy/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.394525 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-677d5bb784-9vt2h_15009458-5da1-43c5-8e7d-9341a8794532/kube-rbac-proxy/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.697912 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-c968bb45-g2828_57067c6a-4af7-4b0f-93cf-ac544f00a239/kube-rbac-proxy/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.707333 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-677d5bb784-9vt2h_15009458-5da1-43c5-8e7d-9341a8794532/operator/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.843668 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xzqmb_8a6a9e34-892e-4580-8cde-73d82b0c65f2/registry-server/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.980946 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-zcr8x_5ecfad04-5c99-47dd-b1f5-782b0aea4061/kube-rbac-proxy/0.log" Oct 05 10:23:32 crc kubenswrapper[4935]: I1005 10:23:32.994071 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-c968bb45-g2828_57067c6a-4af7-4b0f-93cf-ac544f00a239/manager/0.log" Oct 05 10:23:33 crc kubenswrapper[4935]: I1005 10:23:33.104408 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-zcr8x_5ecfad04-5c99-47dd-b1f5-782b0aea4061/manager/0.log" Oct 05 10:23:33 crc kubenswrapper[4935]: I1005 10:23:33.227808 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-z6fkt_98df2e22-67cc-485b-856d-ce3e358b0bfd/operator/0.log" Oct 05 10:23:33 crc kubenswrapper[4935]: I1005 10:23:33.344747 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-8xv6f_8ca55603-341c-4a88-aaec-34ce00cec8fb/kube-rbac-proxy/0.log" Oct 05 10:23:33 crc kubenswrapper[4935]: I1005 10:23:33.455122 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-8xv6f_8ca55603-341c-4a88-aaec-34ce00cec8fb/manager/0.log" Oct 05 10:23:33 crc kubenswrapper[4935]: I1005 10:23:33.511380 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-xkph9_446adb3e-02c1-4aa1-af8b-a038b43a3a61/kube-rbac-proxy/0.log" Oct 05 10:23:33 crc kubenswrapper[4935]: I1005 10:23:33.800785 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-hg5ns_8a31c4d9-3e39-4c59-ac21-f202677474ee/manager/0.log" Oct 05 10:23:33 crc kubenswrapper[4935]: I1005 10:23:33.814336 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-hg5ns_8a31c4d9-3e39-4c59-ac21-f202677474ee/kube-rbac-proxy/0.log" Oct 05 10:23:33 crc kubenswrapper[4935]: I1005 10:23:33.876966 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-xkph9_446adb3e-02c1-4aa1-af8b-a038b43a3a61/manager/0.log" Oct 05 10:23:34 crc kubenswrapper[4935]: I1005 10:23:34.007569 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-rtvd8_57af99f4-85fe-426a-a608-6c46d72cc16f/manager/0.log" Oct 05 10:23:34 crc kubenswrapper[4935]: I1005 10:23:34.070103 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-rtvd8_57af99f4-85fe-426a-a608-6c46d72cc16f/kube-rbac-proxy/0.log" Oct 05 10:23:34 crc kubenswrapper[4935]: I1005 10:23:34.992945 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7cfc658b9-vckkj_3ca2bfab-d801-4cbf-bb97-c0eff17db807/manager/0.log" Oct 05 10:23:37 crc kubenswrapper[4935]: I1005 10:23:37.778354 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:23:37 crc kubenswrapper[4935]: E1005 10:23:37.779100 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:23:49 crc kubenswrapper[4935]: I1005 10:23:49.777641 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:23:49 crc kubenswrapper[4935]: E1005 10:23:49.778443 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:23:52 crc kubenswrapper[4935]: I1005 10:23:52.026027 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zp4tc_c33e835e-eefc-4c71-84ba-a7e6784b17bd/control-plane-machine-set-operator/0.log" Oct 05 10:23:52 crc kubenswrapper[4935]: I1005 10:23:52.179202 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rphcb_4064a1c6-00f1-4c14-9ce9-42308f63069f/kube-rbac-proxy/0.log" Oct 05 10:23:52 crc kubenswrapper[4935]: I1005 10:23:52.240427 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rphcb_4064a1c6-00f1-4c14-9ce9-42308f63069f/machine-api-operator/0.log" Oct 05 10:24:02 crc kubenswrapper[4935]: I1005 10:24:02.777461 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:24:02 crc kubenswrapper[4935]: E1005 10:24:02.778517 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:24:05 crc kubenswrapper[4935]: I1005 10:24:05.237771 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-82msx_171c85a1-08f1-481c-85fd-e60317147ad2/cert-manager-controller/0.log" Oct 05 10:24:05 crc kubenswrapper[4935]: I1005 10:24:05.440691 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-rj6tr_0622bcbf-485e-48f2-8d5f-a359620318c4/cert-manager-webhook/0.log" Oct 05 10:24:05 crc kubenswrapper[4935]: I1005 10:24:05.455747 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-c6tlk_42349445-26e8-4b37-b947-9fe64536e3ab/cert-manager-cainjector/0.log" Oct 05 10:24:15 crc kubenswrapper[4935]: I1005 10:24:15.778199 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:24:15 crc kubenswrapper[4935]: E1005 10:24:15.779350 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:24:18 crc kubenswrapper[4935]: I1005 10:24:18.101741 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-bl5mm_406136d1-87c2-4e34-9941-38d84f0aaff4/nmstate-console-plugin/0.log" Oct 05 10:24:18 crc kubenswrapper[4935]: I1005 10:24:18.318699 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qxgvc_7d9b8112-d308-439e-838d-410a3bc6bf18/nmstate-handler/0.log" Oct 05 10:24:18 crc kubenswrapper[4935]: I1005 10:24:18.344859 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-blpj9_2a702a6b-c2f1-46ae-b6da-a99ff519a83f/kube-rbac-proxy/0.log" Oct 05 10:24:18 crc kubenswrapper[4935]: I1005 10:24:18.375964 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-blpj9_2a702a6b-c2f1-46ae-b6da-a99ff519a83f/nmstate-metrics/0.log" Oct 05 10:24:18 crc kubenswrapper[4935]: I1005 10:24:18.556424 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-x2q48_7f754f36-a0e0-4a97-b03d-f8d5d69f4b00/nmstate-operator/0.log" Oct 05 10:24:18 crc kubenswrapper[4935]: I1005 10:24:18.662076 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-n8448_66a68296-c0ef-463c-8d6f-e6a970833180/nmstate-webhook/0.log" Oct 05 10:24:30 crc kubenswrapper[4935]: I1005 10:24:30.778746 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:24:30 crc kubenswrapper[4935]: E1005 10:24:30.780151 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:24:34 crc kubenswrapper[4935]: I1005 10:24:34.439361 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-4zp2r_a058bfb5-58b4-4ac1-885e-2e2748d2c4ee/kube-rbac-proxy/0.log" Oct 05 10:24:34 crc kubenswrapper[4935]: I1005 10:24:34.639796 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-wfhfw_2875ff58-eeeb-477b-880f-efbfcf23ea8d/frr-k8s-webhook-server/0.log" Oct 05 10:24:34 crc kubenswrapper[4935]: I1005 10:24:34.877866 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-frr-files/0.log" Oct 05 10:24:34 crc kubenswrapper[4935]: I1005 10:24:34.928563 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-4zp2r_a058bfb5-58b4-4ac1-885e-2e2748d2c4ee/controller/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.134817 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-metrics/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.152786 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-frr-files/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.156090 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-reloader/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.163200 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-reloader/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.416101 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-reloader/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.418490 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-metrics/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.457501 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-frr-files/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.476441 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-metrics/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.659265 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-frr-files/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.665547 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-reloader/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.676502 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/cp-metrics/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.734752 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/controller/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.848837 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/frr-metrics/0.log" Oct 05 10:24:35 crc kubenswrapper[4935]: I1005 10:24:35.876260 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/kube-rbac-proxy/0.log" Oct 05 10:24:36 crc kubenswrapper[4935]: I1005 10:24:36.013114 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/kube-rbac-proxy-frr/0.log" Oct 05 10:24:36 crc kubenswrapper[4935]: I1005 10:24:36.082765 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/reloader/0.log" Oct 05 10:24:36 crc kubenswrapper[4935]: I1005 10:24:36.295160 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-944fbc97-lnpwl_efba8319-0b0d-415d-bb5d-7d1381c85524/manager/0.log" Oct 05 10:24:36 crc kubenswrapper[4935]: I1005 10:24:36.389074 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-659857b8df-5kfhr_7c48384d-5634-4320-b93b-b82d3eaf91ca/webhook-server/0.log" Oct 05 10:24:36 crc kubenswrapper[4935]: I1005 10:24:36.482573 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-524b4_07951b31-51bf-45fd-b391-2195e76d97ce/kube-rbac-proxy/0.log" Oct 05 10:24:37 crc kubenswrapper[4935]: I1005 10:24:37.425444 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-524b4_07951b31-51bf-45fd-b391-2195e76d97ce/speaker/0.log" Oct 05 10:24:39 crc kubenswrapper[4935]: I1005 10:24:39.548230 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xs9zt_4fb11792-cce4-4bbf-8df5-420e99305afa/frr/0.log" Oct 05 10:24:45 crc kubenswrapper[4935]: I1005 10:24:45.778230 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:24:45 crc kubenswrapper[4935]: E1005 10:24:45.779283 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:24:51 crc kubenswrapper[4935]: I1005 10:24:51.540025 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k_5e538f9b-2152-463f-b3f9-f4a33f000688/util/0.log" Oct 05 10:24:51 crc kubenswrapper[4935]: I1005 10:24:51.785950 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k_5e538f9b-2152-463f-b3f9-f4a33f000688/util/0.log" Oct 05 10:24:51 crc kubenswrapper[4935]: I1005 10:24:51.837918 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k_5e538f9b-2152-463f-b3f9-f4a33f000688/pull/0.log" Oct 05 10:24:51 crc kubenswrapper[4935]: I1005 10:24:51.841811 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k_5e538f9b-2152-463f-b3f9-f4a33f000688/pull/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.002197 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k_5e538f9b-2152-463f-b3f9-f4a33f000688/util/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.018915 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k_5e538f9b-2152-463f-b3f9-f4a33f000688/extract/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.062326 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69jj86k_5e538f9b-2152-463f-b3f9-f4a33f000688/pull/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.195855 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf_a5685c69-6cce-44df-9252-84962c0cc76a/util/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.574870 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf_a5685c69-6cce-44df-9252-84962c0cc76a/pull/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.645504 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf_a5685c69-6cce-44df-9252-84962c0cc76a/pull/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.675335 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf_a5685c69-6cce-44df-9252-84962c0cc76a/util/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.794982 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf_a5685c69-6cce-44df-9252-84962c0cc76a/pull/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.829976 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf_a5685c69-6cce-44df-9252-84962c0cc76a/util/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.831263 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ntknf_a5685c69-6cce-44df-9252-84962c0cc76a/extract/0.log" Oct 05 10:24:52 crc kubenswrapper[4935]: I1005 10:24:52.973764 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg_4f0ed86d-4848-4b38-b2d0-39a9dceaca1f/util/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.192437 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg_4f0ed86d-4848-4b38-b2d0-39a9dceaca1f/pull/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.208862 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg_4f0ed86d-4848-4b38-b2d0-39a9dceaca1f/util/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.224455 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg_4f0ed86d-4848-4b38-b2d0-39a9dceaca1f/pull/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.412393 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg_4f0ed86d-4848-4b38-b2d0-39a9dceaca1f/util/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.415099 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg_4f0ed86d-4848-4b38-b2d0-39a9dceaca1f/extract/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.425800 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d6sxsg_4f0ed86d-4848-4b38-b2d0-39a9dceaca1f/pull/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.583650 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dqf5x_70bc22d6-6824-4a71-ab0b-b739369accbc/extract-utilities/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.759826 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dqf5x_70bc22d6-6824-4a71-ab0b-b739369accbc/extract-utilities/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.768499 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dqf5x_70bc22d6-6824-4a71-ab0b-b739369accbc/extract-content/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.779749 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dqf5x_70bc22d6-6824-4a71-ab0b-b739369accbc/extract-content/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.922131 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dqf5x_70bc22d6-6824-4a71-ab0b-b739369accbc/extract-content/0.log" Oct 05 10:24:53 crc kubenswrapper[4935]: I1005 10:24:53.927747 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dqf5x_70bc22d6-6824-4a71-ab0b-b739369accbc/extract-utilities/0.log" Oct 05 10:24:54 crc kubenswrapper[4935]: I1005 10:24:54.128865 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s65n7_8007248d-a0fa-4469-97a0-40a224915852/extract-utilities/0.log" Oct 05 10:24:54 crc kubenswrapper[4935]: I1005 10:24:54.396576 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s65n7_8007248d-a0fa-4469-97a0-40a224915852/extract-content/0.log" Oct 05 10:24:54 crc kubenswrapper[4935]: I1005 10:24:54.424786 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s65n7_8007248d-a0fa-4469-97a0-40a224915852/extract-utilities/0.log" Oct 05 10:24:54 crc kubenswrapper[4935]: I1005 10:24:54.497164 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s65n7_8007248d-a0fa-4469-97a0-40a224915852/extract-content/0.log" Oct 05 10:24:54 crc kubenswrapper[4935]: I1005 10:24:54.591785 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s65n7_8007248d-a0fa-4469-97a0-40a224915852/extract-utilities/0.log" Oct 05 10:24:54 crc kubenswrapper[4935]: I1005 10:24:54.624433 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s65n7_8007248d-a0fa-4469-97a0-40a224915852/extract-content/0.log" Oct 05 10:24:54 crc kubenswrapper[4935]: I1005 10:24:54.854121 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz_3553056f-ef6b-407c-8519-255e0f59496c/util/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.095018 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz_3553056f-ef6b-407c-8519-255e0f59496c/util/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.125856 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz_3553056f-ef6b-407c-8519-255e0f59496c/pull/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.148614 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz_3553056f-ef6b-407c-8519-255e0f59496c/pull/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.357803 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz_3553056f-ef6b-407c-8519-255e0f59496c/util/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.421566 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz_3553056f-ef6b-407c-8519-255e0f59496c/pull/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.425398 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8jvxz_3553056f-ef6b-407c-8519-255e0f59496c/extract/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.628355 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-g9n7m_6067d7a8-863b-46f1-97d0-e848aea36ebf/marketplace-operator/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.835342 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cl4vw_9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9/extract-utilities/0.log" Oct 05 10:24:55 crc kubenswrapper[4935]: I1005 10:24:55.993557 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cl4vw_9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9/extract-content/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.047706 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cl4vw_9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9/extract-utilities/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.056778 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dqf5x_70bc22d6-6824-4a71-ab0b-b739369accbc/registry-server/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.116337 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cl4vw_9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9/extract-content/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.349516 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cl4vw_9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9/extract-content/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.380971 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cl4vw_9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9/extract-utilities/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.613503 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gbnlj_d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742/extract-utilities/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.834471 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gbnlj_d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742/extract-content/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.851138 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gbnlj_d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742/extract-content/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.855225 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gbnlj_d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742/extract-utilities/0.log" Oct 05 10:24:56 crc kubenswrapper[4935]: I1005 10:24:56.895376 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s65n7_8007248d-a0fa-4469-97a0-40a224915852/registry-server/0.log" Oct 05 10:24:57 crc kubenswrapper[4935]: I1005 10:24:57.080685 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cl4vw_9a345bf9-50e9-4d7a-8dc8-cf26ccb48ad9/registry-server/0.log" Oct 05 10:24:57 crc kubenswrapper[4935]: I1005 10:24:57.095448 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gbnlj_d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742/extract-utilities/0.log" Oct 05 10:24:57 crc kubenswrapper[4935]: I1005 10:24:57.121528 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gbnlj_d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742/extract-content/0.log" Oct 05 10:24:58 crc kubenswrapper[4935]: I1005 10:24:58.615376 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gbnlj_d3fc5fe9-197b-45d5-aedd-a1f5bc9e8742/registry-server/0.log" Oct 05 10:24:59 crc kubenswrapper[4935]: I1005 10:24:59.777218 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:24:59 crc kubenswrapper[4935]: E1005 10:24:59.777667 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:25:09 crc kubenswrapper[4935]: I1005 10:25:09.698657 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-t8cft_d0952f68-f026-4b91-bf20-41c0f6b3ae09/prometheus-operator/0.log" Oct 05 10:25:09 crc kubenswrapper[4935]: I1005 10:25:09.862999 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-649566fff-bckjm_56891111-0316-4485-9beb-78e3d36385be/prometheus-operator-admission-webhook/0.log" Oct 05 10:25:09 crc kubenswrapper[4935]: I1005 10:25:09.895831 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-649566fff-rxngt_bed510a3-6d84-402d-9c73-22adbae5329f/prometheus-operator-admission-webhook/0.log" Oct 05 10:25:10 crc kubenswrapper[4935]: I1005 10:25:10.056135 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-qhnds_4018d5a3-249d-4dcb-a528-504a9ce098bf/operator/0.log" Oct 05 10:25:10 crc kubenswrapper[4935]: I1005 10:25:10.125683 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-ztdwx_edf2f10d-a59a-4d08-81b2-b86892f02e54/perses-operator/0.log" Oct 05 10:25:11 crc kubenswrapper[4935]: I1005 10:25:11.776687 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:25:11 crc kubenswrapper[4935]: E1005 10:25:11.777197 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:25:16 crc kubenswrapper[4935]: E1005 10:25:16.669666 4935 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.59:49404->38.102.83.59:37063: read tcp 38.102.83.59:49404->38.102.83.59:37063: read: connection reset by peer Oct 05 10:25:16 crc kubenswrapper[4935]: E1005 10:25:16.670282 4935 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.59:49404->38.102.83.59:37063: write tcp 38.102.83.59:49404->38.102.83.59:37063: write: broken pipe Oct 05 10:25:25 crc kubenswrapper[4935]: I1005 10:25:25.777970 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:25:25 crc kubenswrapper[4935]: E1005 10:25:25.778819 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:25:37 crc kubenswrapper[4935]: I1005 10:25:37.777050 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:25:37 crc kubenswrapper[4935]: E1005 10:25:37.778072 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:25:52 crc kubenswrapper[4935]: I1005 10:25:52.780473 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:25:53 crc kubenswrapper[4935]: I1005 10:25:53.884093 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"a875378fbf8805904441f34060956e2fa6f4ed2e2f4d32661b9bf33d8ef43267"} Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.904425 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rplsf"] Oct 05 10:28:03 crc kubenswrapper[4935]: E1005 10:28:03.907669 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c51d88f-00db-488d-860f-7fed6c982492" containerName="extract-utilities" Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.907953 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c51d88f-00db-488d-860f-7fed6c982492" containerName="extract-utilities" Oct 05 10:28:03 crc kubenswrapper[4935]: E1005 10:28:03.908129 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c51d88f-00db-488d-860f-7fed6c982492" containerName="registry-server" Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.908258 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c51d88f-00db-488d-860f-7fed6c982492" containerName="registry-server" Oct 05 10:28:03 crc kubenswrapper[4935]: E1005 10:28:03.908430 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c51d88f-00db-488d-860f-7fed6c982492" containerName="extract-content" Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.908570 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c51d88f-00db-488d-860f-7fed6c982492" containerName="extract-content" Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.909235 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c51d88f-00db-488d-860f-7fed6c982492" containerName="registry-server" Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.915697 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.926870 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rplsf"] Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.997179 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-utilities\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.997469 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42r5c\" (UniqueName: \"kubernetes.io/projected/459876a1-beb5-4236-89b2-2d758396d792-kube-api-access-42r5c\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:03 crc kubenswrapper[4935]: I1005 10:28:03.997628 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-catalog-content\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:04 crc kubenswrapper[4935]: I1005 10:28:04.099226 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-utilities\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:04 crc kubenswrapper[4935]: I1005 10:28:04.099330 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42r5c\" (UniqueName: \"kubernetes.io/projected/459876a1-beb5-4236-89b2-2d758396d792-kube-api-access-42r5c\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:04 crc kubenswrapper[4935]: I1005 10:28:04.099424 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-catalog-content\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:04 crc kubenswrapper[4935]: I1005 10:28:04.099962 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-catalog-content\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:04 crc kubenswrapper[4935]: I1005 10:28:04.100030 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-utilities\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:04 crc kubenswrapper[4935]: I1005 10:28:04.122513 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42r5c\" (UniqueName: \"kubernetes.io/projected/459876a1-beb5-4236-89b2-2d758396d792-kube-api-access-42r5c\") pod \"community-operators-rplsf\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:04 crc kubenswrapper[4935]: I1005 10:28:04.252567 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:04 crc kubenswrapper[4935]: I1005 10:28:04.808449 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rplsf"] Oct 05 10:28:05 crc kubenswrapper[4935]: I1005 10:28:05.787771 4935 generic.go:334] "Generic (PLEG): container finished" podID="459876a1-beb5-4236-89b2-2d758396d792" containerID="ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335" exitCode=0 Oct 05 10:28:05 crc kubenswrapper[4935]: I1005 10:28:05.788203 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rplsf" event={"ID":"459876a1-beb5-4236-89b2-2d758396d792","Type":"ContainerDied","Data":"ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335"} Oct 05 10:28:05 crc kubenswrapper[4935]: I1005 10:28:05.788242 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rplsf" event={"ID":"459876a1-beb5-4236-89b2-2d758396d792","Type":"ContainerStarted","Data":"73e42dda719e28d4757ee1e345ac06556fa9e5d627a449dd5977ee8661392271"} Oct 05 10:28:05 crc kubenswrapper[4935]: I1005 10:28:05.790217 4935 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 10:28:06 crc kubenswrapper[4935]: I1005 10:28:06.527977 4935 scope.go:117] "RemoveContainer" containerID="ab073a62fa6eb6a7112f3dc05222a247726e508dd89db47c4d7bd949af99495c" Oct 05 10:28:06 crc kubenswrapper[4935]: I1005 10:28:06.799354 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rplsf" event={"ID":"459876a1-beb5-4236-89b2-2d758396d792","Type":"ContainerStarted","Data":"2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82"} Oct 05 10:28:08 crc kubenswrapper[4935]: I1005 10:28:08.827936 4935 generic.go:334] "Generic (PLEG): container finished" podID="459876a1-beb5-4236-89b2-2d758396d792" containerID="2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82" exitCode=0 Oct 05 10:28:08 crc kubenswrapper[4935]: I1005 10:28:08.828024 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rplsf" event={"ID":"459876a1-beb5-4236-89b2-2d758396d792","Type":"ContainerDied","Data":"2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82"} Oct 05 10:28:09 crc kubenswrapper[4935]: I1005 10:28:09.850194 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rplsf" event={"ID":"459876a1-beb5-4236-89b2-2d758396d792","Type":"ContainerStarted","Data":"24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1"} Oct 05 10:28:14 crc kubenswrapper[4935]: I1005 10:28:14.253046 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:14 crc kubenswrapper[4935]: I1005 10:28:14.260321 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:14 crc kubenswrapper[4935]: I1005 10:28:14.289697 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:28:14 crc kubenswrapper[4935]: I1005 10:28:14.290135 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:28:14 crc kubenswrapper[4935]: I1005 10:28:14.341885 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:14 crc kubenswrapper[4935]: I1005 10:28:14.373148 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rplsf" podStartSLOduration=7.919397826 podStartE2EDuration="11.373127606s" podCreationTimestamp="2025-10-05 10:28:03 +0000 UTC" firstStartedPulling="2025-10-05 10:28:05.789847263 +0000 UTC m=+12919.672473733" lastFinishedPulling="2025-10-05 10:28:09.243577033 +0000 UTC m=+12923.126203513" observedRunningTime="2025-10-05 10:28:09.903384583 +0000 UTC m=+12923.786011043" watchObservedRunningTime="2025-10-05 10:28:14.373127606 +0000 UTC m=+12928.255754066" Oct 05 10:28:14 crc kubenswrapper[4935]: I1005 10:28:14.997061 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:15 crc kubenswrapper[4935]: I1005 10:28:15.062591 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rplsf"] Oct 05 10:28:16 crc kubenswrapper[4935]: I1005 10:28:16.961765 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rplsf" podUID="459876a1-beb5-4236-89b2-2d758396d792" containerName="registry-server" containerID="cri-o://24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1" gracePeriod=2 Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.631049 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.810166 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-catalog-content\") pod \"459876a1-beb5-4236-89b2-2d758396d792\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.810424 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42r5c\" (UniqueName: \"kubernetes.io/projected/459876a1-beb5-4236-89b2-2d758396d792-kube-api-access-42r5c\") pod \"459876a1-beb5-4236-89b2-2d758396d792\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.810605 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-utilities\") pod \"459876a1-beb5-4236-89b2-2d758396d792\" (UID: \"459876a1-beb5-4236-89b2-2d758396d792\") " Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.812363 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-utilities" (OuterVolumeSpecName: "utilities") pod "459876a1-beb5-4236-89b2-2d758396d792" (UID: "459876a1-beb5-4236-89b2-2d758396d792"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.822498 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/459876a1-beb5-4236-89b2-2d758396d792-kube-api-access-42r5c" (OuterVolumeSpecName: "kube-api-access-42r5c") pod "459876a1-beb5-4236-89b2-2d758396d792" (UID: "459876a1-beb5-4236-89b2-2d758396d792"). InnerVolumeSpecName "kube-api-access-42r5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.863875 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "459876a1-beb5-4236-89b2-2d758396d792" (UID: "459876a1-beb5-4236-89b2-2d758396d792"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.914092 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42r5c\" (UniqueName: \"kubernetes.io/projected/459876a1-beb5-4236-89b2-2d758396d792-kube-api-access-42r5c\") on node \"crc\" DevicePath \"\"" Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.914148 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.914167 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459876a1-beb5-4236-89b2-2d758396d792-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.980746 4935 generic.go:334] "Generic (PLEG): container finished" podID="459876a1-beb5-4236-89b2-2d758396d792" containerID="24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1" exitCode=0 Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.981019 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rplsf" event={"ID":"459876a1-beb5-4236-89b2-2d758396d792","Type":"ContainerDied","Data":"24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1"} Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.981047 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rplsf" event={"ID":"459876a1-beb5-4236-89b2-2d758396d792","Type":"ContainerDied","Data":"73e42dda719e28d4757ee1e345ac06556fa9e5d627a449dd5977ee8661392271"} Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.981064 4935 scope.go:117] "RemoveContainer" containerID="24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1" Oct 05 10:28:17 crc kubenswrapper[4935]: I1005 10:28:17.981198 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rplsf" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.037546 4935 scope.go:117] "RemoveContainer" containerID="2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.043396 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rplsf"] Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.060379 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rplsf"] Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.075226 4935 scope.go:117] "RemoveContainer" containerID="ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.133600 4935 scope.go:117] "RemoveContainer" containerID="24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1" Oct 05 10:28:18 crc kubenswrapper[4935]: E1005 10:28:18.134231 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1\": container with ID starting with 24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1 not found: ID does not exist" containerID="24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.134355 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1"} err="failed to get container status \"24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1\": rpc error: code = NotFound desc = could not find container \"24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1\": container with ID starting with 24ac2aaf810ded379907a9801da1fa8ceab731bf1c3a3106d2e55360333819d1 not found: ID does not exist" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.134399 4935 scope.go:117] "RemoveContainer" containerID="2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82" Oct 05 10:28:18 crc kubenswrapper[4935]: E1005 10:28:18.135201 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82\": container with ID starting with 2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82 not found: ID does not exist" containerID="2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.135257 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82"} err="failed to get container status \"2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82\": rpc error: code = NotFound desc = could not find container \"2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82\": container with ID starting with 2ac99882f40b8fe20e521629af50f3d65a36ca2398ec39d974d15a0a3a90cf82 not found: ID does not exist" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.135273 4935 scope.go:117] "RemoveContainer" containerID="ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335" Oct 05 10:28:18 crc kubenswrapper[4935]: E1005 10:28:18.135705 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335\": container with ID starting with ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335 not found: ID does not exist" containerID="ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.135743 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335"} err="failed to get container status \"ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335\": rpc error: code = NotFound desc = could not find container \"ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335\": container with ID starting with ef44ea624fbac34a5f8a6ad4409aa492b32e628b347b837f053b560873bf0335 not found: ID does not exist" Oct 05 10:28:18 crc kubenswrapper[4935]: I1005 10:28:18.800836 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="459876a1-beb5-4236-89b2-2d758396d792" path="/var/lib/kubelet/pods/459876a1-beb5-4236-89b2-2d758396d792/volumes" Oct 05 10:28:44 crc kubenswrapper[4935]: I1005 10:28:44.289562 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:28:44 crc kubenswrapper[4935]: I1005 10:28:44.290379 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.290169 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.291259 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.291333 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.293087 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a875378fbf8805904441f34060956e2fa6f4ed2e2f4d32661b9bf33d8ef43267"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.293228 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://a875378fbf8805904441f34060956e2fa6f4ed2e2f4d32661b9bf33d8ef43267" gracePeriod=600 Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.882458 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="a875378fbf8805904441f34060956e2fa6f4ed2e2f4d32661b9bf33d8ef43267" exitCode=0 Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.882502 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"a875378fbf8805904441f34060956e2fa6f4ed2e2f4d32661b9bf33d8ef43267"} Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.883007 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerStarted","Data":"3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2"} Oct 05 10:29:14 crc kubenswrapper[4935]: I1005 10:29:14.883042 4935 scope.go:117] "RemoveContainer" containerID="88164f5afb592088b89b2a8504ba3e3b042c0ff3fd2f61eca6d8034c74bb7f24" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.208079 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk"] Oct 05 10:30:00 crc kubenswrapper[4935]: E1005 10:30:00.209164 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459876a1-beb5-4236-89b2-2d758396d792" containerName="extract-content" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.209184 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="459876a1-beb5-4236-89b2-2d758396d792" containerName="extract-content" Oct 05 10:30:00 crc kubenswrapper[4935]: E1005 10:30:00.209214 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459876a1-beb5-4236-89b2-2d758396d792" containerName="extract-utilities" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.209223 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="459876a1-beb5-4236-89b2-2d758396d792" containerName="extract-utilities" Oct 05 10:30:00 crc kubenswrapper[4935]: E1005 10:30:00.209249 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459876a1-beb5-4236-89b2-2d758396d792" containerName="registry-server" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.209258 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="459876a1-beb5-4236-89b2-2d758396d792" containerName="registry-server" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.209502 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="459876a1-beb5-4236-89b2-2d758396d792" containerName="registry-server" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.210362 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.216172 4935 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.216258 4935 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.228423 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk"] Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.265560 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-secret-volume\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.265717 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wdwq\" (UniqueName: \"kubernetes.io/projected/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-kube-api-access-2wdwq\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.266459 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-config-volume\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.378870 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-config-volume\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.383035 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-secret-volume\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.383254 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wdwq\" (UniqueName: \"kubernetes.io/projected/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-kube-api-access-2wdwq\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.387317 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-config-volume\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.416290 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-secret-volume\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.435561 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wdwq\" (UniqueName: \"kubernetes.io/projected/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-kube-api-access-2wdwq\") pod \"collect-profiles-29327670-wdfbk\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:00 crc kubenswrapper[4935]: I1005 10:30:00.533452 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:01 crc kubenswrapper[4935]: I1005 10:30:01.006693 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk"] Oct 05 10:30:01 crc kubenswrapper[4935]: I1005 10:30:01.601942 4935 generic.go:334] "Generic (PLEG): container finished" podID="0b410aee-eeae-4c11-8c8b-ea4798e6ed27" containerID="524e75d3d457c0975e85ae7baeb6f1ba5927fdf34e4a7363beeb76df701b5ea9" exitCode=0 Oct 05 10:30:01 crc kubenswrapper[4935]: I1005 10:30:01.603711 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" event={"ID":"0b410aee-eeae-4c11-8c8b-ea4798e6ed27","Type":"ContainerDied","Data":"524e75d3d457c0975e85ae7baeb6f1ba5927fdf34e4a7363beeb76df701b5ea9"} Oct 05 10:30:01 crc kubenswrapper[4935]: I1005 10:30:01.604563 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" event={"ID":"0b410aee-eeae-4c11-8c8b-ea4798e6ed27","Type":"ContainerStarted","Data":"1fe8afebbcbd0de885647b1df5d4cb02f5e7b65a891eccbc344d68703d9fd1e7"} Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.112334 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.148713 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-config-volume\") pod \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.148966 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wdwq\" (UniqueName: \"kubernetes.io/projected/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-kube-api-access-2wdwq\") pod \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.148998 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-secret-volume\") pod \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\" (UID: \"0b410aee-eeae-4c11-8c8b-ea4798e6ed27\") " Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.150548 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-config-volume" (OuterVolumeSpecName: "config-volume") pod "0b410aee-eeae-4c11-8c8b-ea4798e6ed27" (UID: "0b410aee-eeae-4c11-8c8b-ea4798e6ed27"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.156282 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0b410aee-eeae-4c11-8c8b-ea4798e6ed27" (UID: "0b410aee-eeae-4c11-8c8b-ea4798e6ed27"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.157340 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-kube-api-access-2wdwq" (OuterVolumeSpecName: "kube-api-access-2wdwq") pod "0b410aee-eeae-4c11-8c8b-ea4798e6ed27" (UID: "0b410aee-eeae-4c11-8c8b-ea4798e6ed27"). InnerVolumeSpecName "kube-api-access-2wdwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.250875 4935 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.250919 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wdwq\" (UniqueName: \"kubernetes.io/projected/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-kube-api-access-2wdwq\") on node \"crc\" DevicePath \"\"" Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.250932 4935 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b410aee-eeae-4c11-8c8b-ea4798e6ed27-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.639919 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" event={"ID":"0b410aee-eeae-4c11-8c8b-ea4798e6ed27","Type":"ContainerDied","Data":"1fe8afebbcbd0de885647b1df5d4cb02f5e7b65a891eccbc344d68703d9fd1e7"} Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.640492 4935 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fe8afebbcbd0de885647b1df5d4cb02f5e7b65a891eccbc344d68703d9fd1e7" Oct 05 10:30:03 crc kubenswrapper[4935]: I1005 10:30:03.640064 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327670-wdfbk" Oct 05 10:30:04 crc kubenswrapper[4935]: I1005 10:30:04.237138 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq"] Oct 05 10:30:04 crc kubenswrapper[4935]: I1005 10:30:04.249560 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327625-46bsq"] Oct 05 10:30:04 crc kubenswrapper[4935]: I1005 10:30:04.658495 4935 generic.go:334] "Generic (PLEG): container finished" podID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerID="e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77" exitCode=0 Oct 05 10:30:04 crc kubenswrapper[4935]: I1005 10:30:04.658555 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9m5xl/must-gather-btd99" event={"ID":"94e6a683-891d-4846-987a-42fc12fe8fb5","Type":"ContainerDied","Data":"e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77"} Oct 05 10:30:04 crc kubenswrapper[4935]: I1005 10:30:04.659519 4935 scope.go:117] "RemoveContainer" containerID="e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77" Oct 05 10:30:04 crc kubenswrapper[4935]: I1005 10:30:04.795992 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="053eb908-340c-49d3-9ce3-0b126f4f3d44" path="/var/lib/kubelet/pods/053eb908-340c-49d3-9ce3-0b126f4f3d44/volumes" Oct 05 10:30:05 crc kubenswrapper[4935]: I1005 10:30:05.304761 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9m5xl_must-gather-btd99_94e6a683-891d-4846-987a-42fc12fe8fb5/gather/0.log" Oct 05 10:30:06 crc kubenswrapper[4935]: I1005 10:30:06.630314 4935 scope.go:117] "RemoveContainer" containerID="afd352ee33ce9b0cd1f0b705e78d166423b7730ac592ca91ed33d77052c5335b" Oct 05 10:30:16 crc kubenswrapper[4935]: I1005 10:30:16.763663 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9m5xl/must-gather-btd99"] Oct 05 10:30:16 crc kubenswrapper[4935]: I1005 10:30:16.765108 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-9m5xl/must-gather-btd99" podUID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerName="copy" containerID="cri-o://42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c" gracePeriod=2 Oct 05 10:30:16 crc kubenswrapper[4935]: I1005 10:30:16.773995 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9m5xl/must-gather-btd99"] Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.380742 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9m5xl_must-gather-btd99_94e6a683-891d-4846-987a-42fc12fe8fb5/copy/0.log" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.382030 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.472674 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/94e6a683-891d-4846-987a-42fc12fe8fb5-must-gather-output\") pod \"94e6a683-891d-4846-987a-42fc12fe8fb5\" (UID: \"94e6a683-891d-4846-987a-42fc12fe8fb5\") " Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.473200 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2flq6\" (UniqueName: \"kubernetes.io/projected/94e6a683-891d-4846-987a-42fc12fe8fb5-kube-api-access-2flq6\") pod \"94e6a683-891d-4846-987a-42fc12fe8fb5\" (UID: \"94e6a683-891d-4846-987a-42fc12fe8fb5\") " Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.496620 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e6a683-891d-4846-987a-42fc12fe8fb5-kube-api-access-2flq6" (OuterVolumeSpecName: "kube-api-access-2flq6") pod "94e6a683-891d-4846-987a-42fc12fe8fb5" (UID: "94e6a683-891d-4846-987a-42fc12fe8fb5"). InnerVolumeSpecName "kube-api-access-2flq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.576388 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2flq6\" (UniqueName: \"kubernetes.io/projected/94e6a683-891d-4846-987a-42fc12fe8fb5-kube-api-access-2flq6\") on node \"crc\" DevicePath \"\"" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.890701 4935 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9m5xl_must-gather-btd99_94e6a683-891d-4846-987a-42fc12fe8fb5/copy/0.log" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.891254 4935 generic.go:334] "Generic (PLEG): container finished" podID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerID="42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c" exitCode=143 Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.891336 4935 scope.go:117] "RemoveContainer" containerID="42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.891549 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9m5xl/must-gather-btd99" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.912218 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94e6a683-891d-4846-987a-42fc12fe8fb5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "94e6a683-891d-4846-987a-42fc12fe8fb5" (UID: "94e6a683-891d-4846-987a-42fc12fe8fb5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.924033 4935 scope.go:117] "RemoveContainer" containerID="e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77" Oct 05 10:30:17 crc kubenswrapper[4935]: I1005 10:30:17.988538 4935 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/94e6a683-891d-4846-987a-42fc12fe8fb5-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 05 10:30:18 crc kubenswrapper[4935]: I1005 10:30:18.044132 4935 scope.go:117] "RemoveContainer" containerID="42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c" Oct 05 10:30:18 crc kubenswrapper[4935]: E1005 10:30:18.044765 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c\": container with ID starting with 42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c not found: ID does not exist" containerID="42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c" Oct 05 10:30:18 crc kubenswrapper[4935]: I1005 10:30:18.044833 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c"} err="failed to get container status \"42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c\": rpc error: code = NotFound desc = could not find container \"42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c\": container with ID starting with 42ddebb62c1b132de560095bd09ce66b2d349c4c67a2333c0241ea7589c4532c not found: ID does not exist" Oct 05 10:30:18 crc kubenswrapper[4935]: I1005 10:30:18.044868 4935 scope.go:117] "RemoveContainer" containerID="e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77" Oct 05 10:30:18 crc kubenswrapper[4935]: E1005 10:30:18.045407 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77\": container with ID starting with e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77 not found: ID does not exist" containerID="e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77" Oct 05 10:30:18 crc kubenswrapper[4935]: I1005 10:30:18.045477 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77"} err="failed to get container status \"e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77\": rpc error: code = NotFound desc = could not find container \"e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77\": container with ID starting with e58cd18b3eb447d744dff760418865371b540c3576de1fc498e95ae147aa7a77 not found: ID does not exist" Oct 05 10:30:18 crc kubenswrapper[4935]: I1005 10:30:18.803875 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94e6a683-891d-4846-987a-42fc12fe8fb5" path="/var/lib/kubelet/pods/94e6a683-891d-4846-987a-42fc12fe8fb5/volumes" Oct 05 10:31:14 crc kubenswrapper[4935]: I1005 10:31:14.289151 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:31:14 crc kubenswrapper[4935]: I1005 10:31:14.291261 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:31:44 crc kubenswrapper[4935]: I1005 10:31:44.291801 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:31:44 crc kubenswrapper[4935]: I1005 10:31:44.292744 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.124148 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b6zqv"] Oct 05 10:31:48 crc kubenswrapper[4935]: E1005 10:31:48.125046 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerName="gather" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.125059 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerName="gather" Oct 05 10:31:48 crc kubenswrapper[4935]: E1005 10:31:48.125072 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerName="copy" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.125079 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerName="copy" Oct 05 10:31:48 crc kubenswrapper[4935]: E1005 10:31:48.125088 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b410aee-eeae-4c11-8c8b-ea4798e6ed27" containerName="collect-profiles" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.125095 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b410aee-eeae-4c11-8c8b-ea4798e6ed27" containerName="collect-profiles" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.125301 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerName="copy" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.125310 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b410aee-eeae-4c11-8c8b-ea4798e6ed27" containerName="collect-profiles" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.125335 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e6a683-891d-4846-987a-42fc12fe8fb5" containerName="gather" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.126780 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.152166 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b6zqv"] Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.229817 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-catalog-content\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.229912 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-utilities\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.230167 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg7cd\" (UniqueName: \"kubernetes.io/projected/93b980aa-422f-47e7-bb64-8a964bc0ac95-kube-api-access-wg7cd\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.332119 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-catalog-content\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.332196 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-utilities\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.332261 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg7cd\" (UniqueName: \"kubernetes.io/projected/93b980aa-422f-47e7-bb64-8a964bc0ac95-kube-api-access-wg7cd\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.333962 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-catalog-content\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.334028 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-utilities\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.359857 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg7cd\" (UniqueName: \"kubernetes.io/projected/93b980aa-422f-47e7-bb64-8a964bc0ac95-kube-api-access-wg7cd\") pod \"certified-operators-b6zqv\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.453220 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:48 crc kubenswrapper[4935]: I1005 10:31:48.948444 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b6zqv"] Oct 05 10:31:49 crc kubenswrapper[4935]: I1005 10:31:49.175781 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b6zqv" event={"ID":"93b980aa-422f-47e7-bb64-8a964bc0ac95","Type":"ContainerStarted","Data":"a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089"} Oct 05 10:31:49 crc kubenswrapper[4935]: I1005 10:31:49.176123 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b6zqv" event={"ID":"93b980aa-422f-47e7-bb64-8a964bc0ac95","Type":"ContainerStarted","Data":"122a412e05c562ea0efbdc0c8fcd44d1a94bd073de00e0440595d0830b07ac0e"} Oct 05 10:31:50 crc kubenswrapper[4935]: I1005 10:31:50.189039 4935 generic.go:334] "Generic (PLEG): container finished" podID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerID="a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089" exitCode=0 Oct 05 10:31:50 crc kubenswrapper[4935]: I1005 10:31:50.189141 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b6zqv" event={"ID":"93b980aa-422f-47e7-bb64-8a964bc0ac95","Type":"ContainerDied","Data":"a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089"} Oct 05 10:31:52 crc kubenswrapper[4935]: I1005 10:31:52.216540 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b6zqv" event={"ID":"93b980aa-422f-47e7-bb64-8a964bc0ac95","Type":"ContainerStarted","Data":"6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa"} Oct 05 10:31:53 crc kubenswrapper[4935]: I1005 10:31:53.235739 4935 generic.go:334] "Generic (PLEG): container finished" podID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerID="6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa" exitCode=0 Oct 05 10:31:53 crc kubenswrapper[4935]: I1005 10:31:53.236232 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b6zqv" event={"ID":"93b980aa-422f-47e7-bb64-8a964bc0ac95","Type":"ContainerDied","Data":"6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa"} Oct 05 10:31:54 crc kubenswrapper[4935]: I1005 10:31:54.249211 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b6zqv" event={"ID":"93b980aa-422f-47e7-bb64-8a964bc0ac95","Type":"ContainerStarted","Data":"f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0"} Oct 05 10:31:54 crc kubenswrapper[4935]: I1005 10:31:54.273410 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b6zqv" podStartSLOduration=2.7853331260000003 podStartE2EDuration="6.273377634s" podCreationTimestamp="2025-10-05 10:31:48 +0000 UTC" firstStartedPulling="2025-10-05 10:31:50.192247682 +0000 UTC m=+13144.074874172" lastFinishedPulling="2025-10-05 10:31:53.68029221 +0000 UTC m=+13147.562918680" observedRunningTime="2025-10-05 10:31:54.268279419 +0000 UTC m=+13148.150905919" watchObservedRunningTime="2025-10-05 10:31:54.273377634 +0000 UTC m=+13148.156004144" Oct 05 10:31:58 crc kubenswrapper[4935]: I1005 10:31:58.456130 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:58 crc kubenswrapper[4935]: I1005 10:31:58.456759 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:58 crc kubenswrapper[4935]: I1005 10:31:58.527680 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:59 crc kubenswrapper[4935]: I1005 10:31:59.393808 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:31:59 crc kubenswrapper[4935]: I1005 10:31:59.470857 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b6zqv"] Oct 05 10:32:01 crc kubenswrapper[4935]: I1005 10:32:01.366823 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b6zqv" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerName="registry-server" containerID="cri-o://f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0" gracePeriod=2 Oct 05 10:32:01 crc kubenswrapper[4935]: I1005 10:32:01.989977 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.072735 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-catalog-content\") pod \"93b980aa-422f-47e7-bb64-8a964bc0ac95\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.072837 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg7cd\" (UniqueName: \"kubernetes.io/projected/93b980aa-422f-47e7-bb64-8a964bc0ac95-kube-api-access-wg7cd\") pod \"93b980aa-422f-47e7-bb64-8a964bc0ac95\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.072907 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-utilities\") pod \"93b980aa-422f-47e7-bb64-8a964bc0ac95\" (UID: \"93b980aa-422f-47e7-bb64-8a964bc0ac95\") " Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.075122 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-utilities" (OuterVolumeSpecName: "utilities") pod "93b980aa-422f-47e7-bb64-8a964bc0ac95" (UID: "93b980aa-422f-47e7-bb64-8a964bc0ac95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.084290 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93b980aa-422f-47e7-bb64-8a964bc0ac95-kube-api-access-wg7cd" (OuterVolumeSpecName: "kube-api-access-wg7cd") pod "93b980aa-422f-47e7-bb64-8a964bc0ac95" (UID: "93b980aa-422f-47e7-bb64-8a964bc0ac95"). InnerVolumeSpecName "kube-api-access-wg7cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.164585 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93b980aa-422f-47e7-bb64-8a964bc0ac95" (UID: "93b980aa-422f-47e7-bb64-8a964bc0ac95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.176062 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.176108 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg7cd\" (UniqueName: \"kubernetes.io/projected/93b980aa-422f-47e7-bb64-8a964bc0ac95-kube-api-access-wg7cd\") on node \"crc\" DevicePath \"\"" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.176122 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b980aa-422f-47e7-bb64-8a964bc0ac95-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.387648 4935 generic.go:334] "Generic (PLEG): container finished" podID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerID="f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0" exitCode=0 Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.387704 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b6zqv" event={"ID":"93b980aa-422f-47e7-bb64-8a964bc0ac95","Type":"ContainerDied","Data":"f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0"} Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.387770 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b6zqv" event={"ID":"93b980aa-422f-47e7-bb64-8a964bc0ac95","Type":"ContainerDied","Data":"122a412e05c562ea0efbdc0c8fcd44d1a94bd073de00e0440595d0830b07ac0e"} Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.387797 4935 scope.go:117] "RemoveContainer" containerID="f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.387719 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b6zqv" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.414510 4935 scope.go:117] "RemoveContainer" containerID="6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.447131 4935 scope.go:117] "RemoveContainer" containerID="a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.470151 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b6zqv"] Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.490076 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b6zqv"] Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.522961 4935 scope.go:117] "RemoveContainer" containerID="f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0" Oct 05 10:32:02 crc kubenswrapper[4935]: E1005 10:32:02.523647 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0\": container with ID starting with f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0 not found: ID does not exist" containerID="f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.523710 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0"} err="failed to get container status \"f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0\": rpc error: code = NotFound desc = could not find container \"f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0\": container with ID starting with f12539ac8d738fae07bf1ceb59de7e23de851e05f185a6e1c35feab0040a66a0 not found: ID does not exist" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.523742 4935 scope.go:117] "RemoveContainer" containerID="6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa" Oct 05 10:32:02 crc kubenswrapper[4935]: E1005 10:32:02.524263 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa\": container with ID starting with 6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa not found: ID does not exist" containerID="6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.524283 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa"} err="failed to get container status \"6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa\": rpc error: code = NotFound desc = could not find container \"6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa\": container with ID starting with 6dd6f424ed43122104d2c3001063411aacb3a8fb2652401c77c91706c02211fa not found: ID does not exist" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.524297 4935 scope.go:117] "RemoveContainer" containerID="a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089" Oct 05 10:32:02 crc kubenswrapper[4935]: E1005 10:32:02.524788 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089\": container with ID starting with a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089 not found: ID does not exist" containerID="a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.524844 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089"} err="failed to get container status \"a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089\": rpc error: code = NotFound desc = could not find container \"a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089\": container with ID starting with a8f587530b3427d5e20f53c628aa48077c29c3c45f7c1e3a5419f6d7fe0ec089 not found: ID does not exist" Oct 05 10:32:02 crc kubenswrapper[4935]: I1005 10:32:02.809463 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" path="/var/lib/kubelet/pods/93b980aa-422f-47e7-bb64-8a964bc0ac95/volumes" Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.289063 4935 patch_prober.go:28] interesting pod/machine-config-daemon-j8cwh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.291983 4935 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.292220 4935 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.294162 4935 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2"} pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.294450 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerName="machine-config-daemon" containerID="cri-o://3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2" gracePeriod=600 Oct 05 10:32:14 crc kubenswrapper[4935]: E1005 10:32:14.434805 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.577846 4935 generic.go:334] "Generic (PLEG): container finished" podID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" containerID="3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2" exitCode=0 Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.577943 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" event={"ID":"62823424-a6c9-4733-b50a-6f3ecb4a55a2","Type":"ContainerDied","Data":"3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2"} Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.578359 4935 scope.go:117] "RemoveContainer" containerID="a875378fbf8805904441f34060956e2fa6f4ed2e2f4d32661b9bf33d8ef43267" Oct 05 10:32:14 crc kubenswrapper[4935]: I1005 10:32:14.579639 4935 scope.go:117] "RemoveContainer" containerID="3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2" Oct 05 10:32:14 crc kubenswrapper[4935]: E1005 10:32:14.580177 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:32:26 crc kubenswrapper[4935]: I1005 10:32:26.790334 4935 scope.go:117] "RemoveContainer" containerID="3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2" Oct 05 10:32:26 crc kubenswrapper[4935]: E1005 10:32:26.791152 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.070539 4935 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w69fc"] Oct 05 10:32:33 crc kubenswrapper[4935]: E1005 10:32:33.072315 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerName="extract-content" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.072341 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerName="extract-content" Oct 05 10:32:33 crc kubenswrapper[4935]: E1005 10:32:33.072399 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerName="registry-server" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.072411 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerName="registry-server" Oct 05 10:32:33 crc kubenswrapper[4935]: E1005 10:32:33.072461 4935 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerName="extract-utilities" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.072474 4935 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerName="extract-utilities" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.072889 4935 memory_manager.go:354] "RemoveStaleState removing state" podUID="93b980aa-422f-47e7-bb64-8a964bc0ac95" containerName="registry-server" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.076179 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.093491 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w69fc"] Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.129674 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-utilities\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.129756 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvzkc\" (UniqueName: \"kubernetes.io/projected/146095f5-7c33-4cd0-8170-583dafdf5659-kube-api-access-zvzkc\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.129936 4935 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-catalog-content\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.232171 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-utilities\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.232243 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvzkc\" (UniqueName: \"kubernetes.io/projected/146095f5-7c33-4cd0-8170-583dafdf5659-kube-api-access-zvzkc\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.232445 4935 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-catalog-content\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.232744 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-utilities\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.233848 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-catalog-content\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.259565 4935 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvzkc\" (UniqueName: \"kubernetes.io/projected/146095f5-7c33-4cd0-8170-583dafdf5659-kube-api-access-zvzkc\") pod \"redhat-operators-w69fc\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.412717 4935 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:33 crc kubenswrapper[4935]: I1005 10:32:33.928231 4935 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w69fc"] Oct 05 10:32:34 crc kubenswrapper[4935]: I1005 10:32:34.854694 4935 generic.go:334] "Generic (PLEG): container finished" podID="146095f5-7c33-4cd0-8170-583dafdf5659" containerID="5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2" exitCode=0 Oct 05 10:32:34 crc kubenswrapper[4935]: I1005 10:32:34.854806 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w69fc" event={"ID":"146095f5-7c33-4cd0-8170-583dafdf5659","Type":"ContainerDied","Data":"5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2"} Oct 05 10:32:34 crc kubenswrapper[4935]: I1005 10:32:34.855159 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w69fc" event={"ID":"146095f5-7c33-4cd0-8170-583dafdf5659","Type":"ContainerStarted","Data":"1e32551f9312d7276d66d1ef15571ce2b0a5354b83d19225932d6cf0d58c8962"} Oct 05 10:32:35 crc kubenswrapper[4935]: I1005 10:32:35.867199 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w69fc" event={"ID":"146095f5-7c33-4cd0-8170-583dafdf5659","Type":"ContainerStarted","Data":"c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b"} Oct 05 10:32:38 crc kubenswrapper[4935]: I1005 10:32:38.907464 4935 generic.go:334] "Generic (PLEG): container finished" podID="146095f5-7c33-4cd0-8170-583dafdf5659" containerID="c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b" exitCode=0 Oct 05 10:32:38 crc kubenswrapper[4935]: I1005 10:32:38.907611 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w69fc" event={"ID":"146095f5-7c33-4cd0-8170-583dafdf5659","Type":"ContainerDied","Data":"c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b"} Oct 05 10:32:40 crc kubenswrapper[4935]: I1005 10:32:40.778660 4935 scope.go:117] "RemoveContainer" containerID="3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2" Oct 05 10:32:40 crc kubenswrapper[4935]: E1005 10:32:40.780178 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:32:40 crc kubenswrapper[4935]: I1005 10:32:40.946557 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w69fc" event={"ID":"146095f5-7c33-4cd0-8170-583dafdf5659","Type":"ContainerStarted","Data":"d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075"} Oct 05 10:32:40 crc kubenswrapper[4935]: I1005 10:32:40.990142 4935 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w69fc" podStartSLOduration=3.501913624 podStartE2EDuration="7.990110214s" podCreationTimestamp="2025-10-05 10:32:33 +0000 UTC" firstStartedPulling="2025-10-05 10:32:34.858232812 +0000 UTC m=+13188.740859282" lastFinishedPulling="2025-10-05 10:32:39.346429372 +0000 UTC m=+13193.229055872" observedRunningTime="2025-10-05 10:32:40.975955249 +0000 UTC m=+13194.858581769" watchObservedRunningTime="2025-10-05 10:32:40.990110214 +0000 UTC m=+13194.872736684" Oct 05 10:32:43 crc kubenswrapper[4935]: I1005 10:32:43.413188 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:43 crc kubenswrapper[4935]: I1005 10:32:43.413279 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:44 crc kubenswrapper[4935]: I1005 10:32:44.469419 4935 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w69fc" podUID="146095f5-7c33-4cd0-8170-583dafdf5659" containerName="registry-server" probeResult="failure" output=< Oct 05 10:32:44 crc kubenswrapper[4935]: timeout: failed to connect service ":50051" within 1s Oct 05 10:32:44 crc kubenswrapper[4935]: > Oct 05 10:32:52 crc kubenswrapper[4935]: I1005 10:32:52.778301 4935 scope.go:117] "RemoveContainer" containerID="3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2" Oct 05 10:32:52 crc kubenswrapper[4935]: E1005 10:32:52.779348 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:32:53 crc kubenswrapper[4935]: I1005 10:32:53.484461 4935 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:53 crc kubenswrapper[4935]: I1005 10:32:53.554377 4935 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:53 crc kubenswrapper[4935]: I1005 10:32:53.749544 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w69fc"] Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.139450 4935 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w69fc" podUID="146095f5-7c33-4cd0-8170-583dafdf5659" containerName="registry-server" containerID="cri-o://d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075" gracePeriod=2 Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.705598 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.805135 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvzkc\" (UniqueName: \"kubernetes.io/projected/146095f5-7c33-4cd0-8170-583dafdf5659-kube-api-access-zvzkc\") pod \"146095f5-7c33-4cd0-8170-583dafdf5659\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.805648 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-utilities\") pod \"146095f5-7c33-4cd0-8170-583dafdf5659\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.805968 4935 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-catalog-content\") pod \"146095f5-7c33-4cd0-8170-583dafdf5659\" (UID: \"146095f5-7c33-4cd0-8170-583dafdf5659\") " Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.807037 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-utilities" (OuterVolumeSpecName: "utilities") pod "146095f5-7c33-4cd0-8170-583dafdf5659" (UID: "146095f5-7c33-4cd0-8170-583dafdf5659"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.818733 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/146095f5-7c33-4cd0-8170-583dafdf5659-kube-api-access-zvzkc" (OuterVolumeSpecName: "kube-api-access-zvzkc") pod "146095f5-7c33-4cd0-8170-583dafdf5659" (UID: "146095f5-7c33-4cd0-8170-583dafdf5659"). InnerVolumeSpecName "kube-api-access-zvzkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.880569 4935 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "146095f5-7c33-4cd0-8170-583dafdf5659" (UID: "146095f5-7c33-4cd0-8170-583dafdf5659"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.911685 4935 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.911774 4935 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/146095f5-7c33-4cd0-8170-583dafdf5659-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 10:32:55 crc kubenswrapper[4935]: I1005 10:32:55.911832 4935 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvzkc\" (UniqueName: \"kubernetes.io/projected/146095f5-7c33-4cd0-8170-583dafdf5659-kube-api-access-zvzkc\") on node \"crc\" DevicePath \"\"" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.156546 4935 generic.go:334] "Generic (PLEG): container finished" podID="146095f5-7c33-4cd0-8170-583dafdf5659" containerID="d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075" exitCode=0 Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.156651 4935 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w69fc" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.156680 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w69fc" event={"ID":"146095f5-7c33-4cd0-8170-583dafdf5659","Type":"ContainerDied","Data":"d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075"} Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.157685 4935 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w69fc" event={"ID":"146095f5-7c33-4cd0-8170-583dafdf5659","Type":"ContainerDied","Data":"1e32551f9312d7276d66d1ef15571ce2b0a5354b83d19225932d6cf0d58c8962"} Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.157732 4935 scope.go:117] "RemoveContainer" containerID="d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.204451 4935 scope.go:117] "RemoveContainer" containerID="c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.207302 4935 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w69fc"] Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.225416 4935 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w69fc"] Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.230291 4935 scope.go:117] "RemoveContainer" containerID="5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.299331 4935 scope.go:117] "RemoveContainer" containerID="d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075" Oct 05 10:32:56 crc kubenswrapper[4935]: E1005 10:32:56.299841 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075\": container with ID starting with d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075 not found: ID does not exist" containerID="d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.299904 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075"} err="failed to get container status \"d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075\": rpc error: code = NotFound desc = could not find container \"d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075\": container with ID starting with d4680f72f3acebea203460e6b3adf3fb5ebe290dee25d116fcb54e00cfbe9075 not found: ID does not exist" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.299935 4935 scope.go:117] "RemoveContainer" containerID="c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b" Oct 05 10:32:56 crc kubenswrapper[4935]: E1005 10:32:56.300324 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b\": container with ID starting with c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b not found: ID does not exist" containerID="c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.300443 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b"} err="failed to get container status \"c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b\": rpc error: code = NotFound desc = could not find container \"c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b\": container with ID starting with c50449f3f8370fc0d0f43e7b04d41d635d32f889bf3813fe864eecb355cd277b not found: ID does not exist" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.300519 4935 scope.go:117] "RemoveContainer" containerID="5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2" Oct 05 10:32:56 crc kubenswrapper[4935]: E1005 10:32:56.300971 4935 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2\": container with ID starting with 5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2 not found: ID does not exist" containerID="5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.301096 4935 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2"} err="failed to get container status \"5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2\": rpc error: code = NotFound desc = could not find container \"5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2\": container with ID starting with 5bff8b945bcc987b6a6bb515801a399621d10f578ae33311019e5ee3346a6bd2 not found: ID does not exist" Oct 05 10:32:56 crc kubenswrapper[4935]: I1005 10:32:56.794794 4935 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="146095f5-7c33-4cd0-8170-583dafdf5659" path="/var/lib/kubelet/pods/146095f5-7c33-4cd0-8170-583dafdf5659/volumes" Oct 05 10:33:05 crc kubenswrapper[4935]: I1005 10:33:05.777047 4935 scope.go:117] "RemoveContainer" containerID="3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2" Oct 05 10:33:05 crc kubenswrapper[4935]: E1005 10:33:05.777943 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" Oct 05 10:33:17 crc kubenswrapper[4935]: I1005 10:33:17.777673 4935 scope.go:117] "RemoveContainer" containerID="3c4feadb8c92a3c9f0a158ddd1849e4b9ee798fe2d4ee09df812852eac2ab3b2" Oct 05 10:33:17 crc kubenswrapper[4935]: E1005 10:33:17.778589 4935 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j8cwh_openshift-machine-config-operator(62823424-a6c9-4733-b50a-6f3ecb4a55a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-j8cwh" podUID="62823424-a6c9-4733-b50a-6f3ecb4a55a2" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070444602024447 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070444603017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070411732016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070411732015455 5ustar corecore